code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm _a : Any = logging.get_logger(__name__) @dataclass class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : List[str] = [ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self , **a__ ): for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: _lowerCAmelCase : Tuple = deprecated_arg[3:] setattr(self , a__ , not kwargs.pop(a__ ) ) logger.warning( F"{deprecated_arg} is depreciated. Please use --no_{positive_arg} or" F" {positive_arg}={kwargs[positive_arg]}" ) _lowerCAmelCase : List[Any] = kwargs.pop("""torchscript""" , self.torchscript ) _lowerCAmelCase : List[str] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) _lowerCAmelCase : List[str] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**a__ ) _UpperCamelCase : bool = field(default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Trace the models using torchscript"} ) _UpperCamelCase : bool = field(default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Print Xla/PyTorch tpu metrics"} ) _UpperCamelCase : str = field( default="O1" , metadata={ "help": ( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. " "See details at https://nvidia.github.io/apex/amp.html" ) } , ) @cached_property def __A ( self ): requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: _lowerCAmelCase : int = torch.device("""cpu""" ) _lowerCAmelCase : Union[str, Any] = 0 elif is_torch_tpu_available(): _lowerCAmelCase : str = xm.xla_device() _lowerCAmelCase : Optional[Any] = 0 else: _lowerCAmelCase : Union[str, Any] = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) _lowerCAmelCase : Optional[Any] = torch.cuda.device_count() return device, n_gpu @property def __A ( self ): return is_torch_tpu_available() and self.tpu @property def __A ( self ): requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def __A ( self ): requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def __A ( self ): requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def __A ( self ): return self.n_gpu > 0
44
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING _a : Dict = logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE_ ) class __A ( SCREAMING_SNAKE_CASE_ ): def __init__( self , *a__ , **a__ ): super().__init__(*a__ , **a__ ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == """tf""" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def __A ( self , a__=None , a__=None , a__=None ): _lowerCAmelCase : List[str] = {} _lowerCAmelCase : Union[str, Any] = {} if prompt is not None: _lowerCAmelCase : List[Any] = prompt if generate_kwargs is not None: _lowerCAmelCase : List[str] = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: _lowerCAmelCase : str = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( """'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter,""" """ please use only one""" ) _lowerCAmelCase : Optional[Any] = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self , a__ , **a__ ): return super().__call__(a__ , **a__ ) def __A ( self , a__ , a__=None ): _lowerCAmelCase : Tuple = load_image(a__ ) if prompt is not None: if not isinstance(a__ , a__ ): raise ValueError( F"Received an invalid text input, got - {type(a__ )} - but expected a single string. " """Note also that one single text can be provided for conditional image to text generation.""" ) _lowerCAmelCase : Optional[int] = self.model.config.model_type if model_type == "git": _lowerCAmelCase : Optional[Any] = self.image_processor(images=a__ , return_tensors=self.framework ) _lowerCAmelCase : List[str] = self.tokenizer(text=a__ , add_special_tokens=a__ ).input_ids _lowerCAmelCase : Union[str, Any] = [self.tokenizer.cls_token_id] + input_ids _lowerCAmelCase : Dict = torch.tensor(a__ ).unsqueeze(0 ) model_inputs.update({"""input_ids""": input_ids} ) elif model_type == "pix2struct": _lowerCAmelCase : Tuple = self.image_processor(images=a__ , header_text=a__ , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation _lowerCAmelCase : Optional[int] = self.image_processor(images=a__ , return_tensors=self.framework ) _lowerCAmelCase : Optional[int] = self.tokenizer(a__ , return_tensors=self.framework ) model_inputs.update(a__ ) else: raise ValueError(F"Model type {model_type} does not support conditional text generation" ) else: _lowerCAmelCase : Any = self.image_processor(images=a__ , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: _lowerCAmelCase : Union[str, Any] = None return model_inputs def __A ( self , a__ , a__=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["""input_ids"""] , a__ ) and all(x is None for x in model_inputs["""input_ids"""] ) ): _lowerCAmelCase : Optional[int] = None if generate_kwargs is None: _lowerCAmelCase : List[str] = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. _lowerCAmelCase : Tuple = model_inputs.pop(self.model.main_input_name ) _lowerCAmelCase : Union[str, Any] = self.model.generate(a__ , **a__ , **a__ ) return model_outputs def __A ( self , a__ ): _lowerCAmelCase : Optional[int] = [] for output_ids in model_outputs: _lowerCAmelCase : Any = { """generated_text""": self.tokenizer.decode( a__ , skip_special_tokens=a__ , ) } records.append(a__ ) return records
44
1
'''simple docstring''' import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self : List[Any] ): _a = 0 @slow def UpperCamelCase__ ( self : Dict ): for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): _a = AutoTokenizer.from_pretrained(__a ) self.assertIsNotNone(__a ) self.assertIsInstance(__a , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(__a ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): _a = AutoTokenizer.from_pretrained(__a ) self.assertIsNotNone(__a ) self.assertIsInstance(__a , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(__a ) , 0 ) def UpperCamelCase__ ( self : List[Any] ): _a = AutoTokenizer.from_pretrained(__a ) self.assertIsInstance(__a , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def UpperCamelCase__ ( self : Union[str, Any] ): _a = AutoTokenizer.from_pretrained(__a ) self.assertIsInstance(__a , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 20 ) def UpperCamelCase__ ( self : List[Any] ): _a = AutoConfig.from_pretrained(__a ) self.assertIsInstance(__a , __a ) # Check that tokenizer_type ≠ model_type _a = AutoTokenizer.from_pretrained(__a , config=__a ) self.assertIsInstance(__a , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def UpperCamelCase__ ( self : str ): with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("./tests/fixtures/vocab.txt" , os.path.join(__a , "vocab.txt" ) ) _a = AutoTokenizer.from_pretrained(__a , tokenizer_type="bert" , use_fast=__a ) self.assertIsInstance(__a , __a ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("./tests/fixtures/vocab.json" , os.path.join(__a , "vocab.json" ) ) shutil.copy("./tests/fixtures/merges.txt" , os.path.join(__a , "merges.txt" ) ) _a = AutoTokenizer.from_pretrained(__a , tokenizer_type="gpt2" , use_fast=__a ) self.assertIsInstance(__a , __a ) @require_tokenizers def UpperCamelCase__ ( self : int ): with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("./tests/fixtures/vocab.txt" , os.path.join(__a , "vocab.txt" ) ) _a = AutoTokenizer.from_pretrained(__a , tokenizer_type="bert" ) self.assertIsInstance(__a , __a ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("./tests/fixtures/vocab.json" , os.path.join(__a , "vocab.json" ) ) shutil.copy("./tests/fixtures/merges.txt" , os.path.join(__a , "merges.txt" ) ) _a = AutoTokenizer.from_pretrained(__a , tokenizer_type="gpt2" ) self.assertIsInstance(__a , __a ) def UpperCamelCase__ ( self : List[str] ): with pytest.raises(__a ): AutoTokenizer.from_pretrained("./" , tokenizer_type="xxx" ) @require_tokenizers def UpperCamelCase__ ( self : str ): for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: _a = tokenizer_class.from_pretrained("wietsedv/bert-base-dutch-cased" ) self.assertIsInstance(__a , (BertTokenizer, BertTokenizerFast) ) if isinstance(__a , __a ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , __a ) else: self.assertEqual(tokenizer.do_lower_case , __a ) self.assertEqual(tokenizer.model_max_length , 5_12 ) @require_tokenizers def UpperCamelCase__ ( self : Optional[int] ): for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( __a , "julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier" , ): _a = tokenizer_class.from_pretrained("julien-c/herlolip-not-exists" ) def UpperCamelCase__ ( self : Optional[Any] ): # tests: https://github.com/huggingface/transformers/pull/13251 # 1. models with `-`, e.g. xlm-roberta -> xlm_roberta # 2. models that don't remap 1-1 from model-name to model file, e.g., openai-gpt -> openai _a = TOKENIZER_MAPPING.values() _a = [] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(__a ) @require_tokenizers def UpperCamelCase__ ( self : str ): self.assertIsInstance(AutoTokenizer.from_pretrained("bert-base-cased" , use_fast=__a ) , __a ) self.assertIsInstance(AutoTokenizer.from_pretrained("bert-base-cased" ) , __a ) @require_tokenizers def UpperCamelCase__ ( self : List[Any] ): _a = AutoTokenizer.from_pretrained("distilbert-base-uncased" , do_lower_case=__a ) _a = "Hello, world. How are you?" _a = tokenizer.tokenize(__a ) self.assertEqual("[UNK]" , tokens[0] ) _a = AutoTokenizer.from_pretrained("microsoft/mpnet-base" , do_lower_case=__a ) _a = tokenizer.tokenize(__a ) self.assertEqual("[UNK]" , tokens[0] ) @require_tokenizers def UpperCamelCase__ ( self : Dict ): _a = AutoTokenizer.from_pretrained("robot-test/dummy-tokenizer-fast-with-model-config" ) self.assertEqual(type(__a ) , __a ) self.assertEqual(tokenizer.model_max_length , 5_12 ) self.assertEqual(tokenizer.vocab_size , 3_00_00 ) self.assertEqual(tokenizer.unk_token , "[UNK]" ) self.assertEqual(tokenizer.padding_side , "right" ) self.assertEqual(tokenizer.truncation_side , "right" ) def UpperCamelCase__ ( self : Optional[Any] ): _a = AutoTokenizer.from_pretrained(__a ) self.assertIsInstance(__a , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a ) _a = AutoTokenizer.from_pretrained(__a ) self.assertIsInstance(__a , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 12 ) def UpperCamelCase__ ( self : Optional[int] ): _a = AutoTokenizer.from_pretrained("ctrl" ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(__a , __a ) def UpperCamelCase__ ( self : List[Any] ): # Check we can load the tokenizer config of an online model. _a = get_tokenizer_config("bert-base-cased" ) _a = config.pop("_commit_hash" , __a ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(__a , {"do_lower_case": False} ) # This model does not have a tokenizer_config so we get back an empty dict. _a = get_tokenizer_config(__a ) self.assertDictEqual(__a , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. _a = AutoTokenizer.from_pretrained(__a ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a ) _a = get_tokenizer_config(__a ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config["tokenizer_class"] , "BertTokenizer" ) def UpperCamelCase__ ( self : str ): try: AutoConfig.register("custom" , __a ) AutoTokenizer.register(__a , slow_tokenizer_class=__a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__a ): AutoTokenizer.register(__a , slow_tokenizer_class=__a ) _a = CustomTokenizer.from_pretrained(__a ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a ) _a = AutoTokenizer.from_pretrained(__a ) self.assertIsInstance(__a , __a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def UpperCamelCase__ ( self : Optional[int] ): try: AutoConfig.register("custom" , __a ) # Can register in two steps AutoTokenizer.register(__a , slow_tokenizer_class=__a ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(__a , fast_tokenizer_class=__a ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( __a , slow_tokenizer_class=__a , fast_tokenizer_class=__a ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__a ): AutoTokenizer.register(__a , fast_tokenizer_class=__a ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: _a = BertTokenizerFast.from_pretrained(__a ) bert_tokenizer.save_pretrained(__a ) _a = CustomTokenizerFast.from_pretrained(__a ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a ) _a = AutoTokenizer.from_pretrained(__a ) self.assertIsInstance(__a , __a ) _a = AutoTokenizer.from_pretrained(__a , use_fast=__a ) self.assertIsInstance(__a , __a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def UpperCamelCase__ ( self : Any ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__a ): _a = AutoTokenizer.from_pretrained("hf-internal-testing/test_dynamic_tokenizer" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__a ): _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a ) _a = AutoTokenizer.from_pretrained("hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a ) _a = AutoTokenizer.from_pretrained(__a , trust_remote_code=__a ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizerFast" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , "NewTokenizerFast" ) # Test we can also load the slow version _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a , use_fast=__a ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a ) _a = AutoTokenizer.from_pretrained(__a , trust_remote_code=__a , use_fast=__a ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , "NewTokenizer" ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , "NewTokenizer" ) @require_tokenizers def UpperCamelCase__ ( self : List[str] ): class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a =False class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a =NewTokenizer __a =False try: AutoConfig.register("custom" , __a ) AutoTokenizer.register(__a , slow_tokenizer_class=__a ) AutoTokenizer.register(__a , fast_tokenizer_class=__a ) # If remote code is not set, the default is to use local _a = AutoTokenizer.from_pretrained("hf-internal-testing/test_dynamic_tokenizer" ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizerFast" ) self.assertFalse(tokenizer.special_attribute_present ) _a = AutoTokenizer.from_pretrained("hf-internal-testing/test_dynamic_tokenizer" , use_fast=__a ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizerFast" ) self.assertFalse(tokenizer.special_attribute_present ) _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a , use_fast=__a ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizerFast" ) self.assertTrue(tokenizer.special_attribute_present ) _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer" , trust_remote_code=__a , use_fast=__a ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def UpperCamelCase__ ( self : List[str] ): _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer_legacy" , trust_remote_code=__a ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizerFast" ) # Test we can also load the slow version _a = AutoTokenizer.from_pretrained( "hf-internal-testing/test_dynamic_tokenizer_legacy" , trust_remote_code=__a , use_fast=__a ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) else: self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) def UpperCamelCase__ ( self : str ): with self.assertRaisesRegex( __a , "bert-base is not a local folder and is not a valid model identifier" ): _a = AutoTokenizer.from_pretrained("bert-base" ) def UpperCamelCase__ ( self : List[str] ): with self.assertRaisesRegex( __a , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): _a = AutoTokenizer.from_pretrained(__a , revision="aaaaaa" ) def UpperCamelCase__ ( self : str ): # Make sure we have cached the tokenizer. _a = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) with RequestCounter() as counter: _a = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
346
'''simple docstring''' from manim import * class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" def UpperCamelCase__ ( self : Dict ): _a = Rectangle(height=0.5 , width=0.5 ) _a = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _a = [mem.copy() for i in range(6 )] _a = [mem.copy() for i in range(6 )] _a = VGroup(*__a ).arrange(__a , buff=0 ) _a = VGroup(*__a ).arrange(__a , buff=0 ) _a = VGroup(__a , __a ).arrange(__a , buff=0 ) _a = Text("CPU" , font_size=24 ) _a = Group(__a , __a ).arrange(__a , buff=0.5 , aligned_edge=__a ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__a ) _a = [mem.copy() for i in range(4 )] _a = VGroup(*__a ).arrange(__a , buff=0 ) _a = Text("GPU" , font_size=24 ) _a = Group(__a , __a ).arrange(__a , buff=0.5 , aligned_edge=__a ) gpu.move_to([-1, -1, 0] ) self.add(__a ) _a = [mem.copy() for i in range(6 )] _a = VGroup(*__a ).arrange(__a , buff=0 ) _a = Text("Model" , font_size=24 ) _a = Group(__a , __a ).arrange(__a , buff=0.5 , aligned_edge=__a ) model.move_to([3, -1.0, 0] ) self.add(__a ) _a = [] for i, rect in enumerate(__a ): rect.set_stroke(__a ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _a = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__a , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__a ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__a , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__a , buff=0.0 ) self.add(__a ) cpu_targs.append(__a ) _a = [mem.copy() for i in range(6 )] _a = VGroup(*__a ).arrange(__a , buff=0 ) _a = Text("Loaded Checkpoint" , font_size=24 ) _a = Group(__a , __a ).arrange(__a , aligned_edge=__a , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _a = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _a = MarkupText( f'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__a , __a ) _a = MarkupText( f'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' , font_size=18 , ) blue_text.next_to(__a , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _a = MarkupText( f'Next, a <i><span fgcolor="{BLUE}">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor="{BLUE}">single shard</span>.' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__a ) , Write(__a ) ) self.play(Write(__a , run_time=1 ) , Create(__a , run_time=1 ) ) _a = [] _a = [] for i, rect in enumerate(__a ): _a = fill.copy().set_fill(__a , opacity=0.7 ) target.move_to(__a ) first_animations.append(GrowFromCenter(__a , run_time=1 ) ) _a = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__a , run_time=1.5 ) ) self.play(*__a ) self.play(*__a ) self.wait()
346
1
import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _lowerCAmelCase ( unittest.TestCase ): def __a ( self ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __a ( self ) -> Any: torch.manual_seed(0 ) lowerCAmelCase_ = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("AttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "AttnUpBlock2D") , ) return model @property def __a ( self ) -> int: torch.manual_seed(0 ) lowerCAmelCase_ = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , cross_attention_dim=10 , ) return model @property def __a ( self ) -> Optional[Any]: torch.manual_seed(0 ) lowerCAmelCase_ = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("DownEncoderBlock2D", "DownEncoderBlock2D") , up_block_types=("UpDecoderBlock2D", "UpDecoderBlock2D") , ) lowerCAmelCase_ = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("AttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "AttnUpBlock2D") , ) return vqvae, unet @slow def __a ( self ) -> int: lowerCAmelCase_ = """cpu""" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) lowerCAmelCase_ = DDPMScheduler() lowerCAmelCase_ = AudioDiffusionPipeline(vqvae=_UpperCamelCase , unet=self.dummy_unet , mel=_UpperCamelCase , scheduler=_UpperCamelCase ) lowerCAmelCase_ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase_ = torch.Generator(device=_UpperCamelCase ).manual_seed(42 ) lowerCAmelCase_ = pipe(generator=_UpperCamelCase , steps=4 ) lowerCAmelCase_ = output.audios[0] lowerCAmelCase_ = output.images[0] lowerCAmelCase_ = torch.Generator(device=_UpperCamelCase ).manual_seed(42 ) lowerCAmelCase_ = pipe(generator=_UpperCamelCase , steps=4 , return_dict=_UpperCamelCase ) lowerCAmelCase_ = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) lowerCAmelCase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowerCAmelCase_ = np.frombuffer(image_from_tuple.tobytes() , dtype="uint8" )[:10] lowerCAmelCase_ = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 lowerCAmelCase_ = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) lowerCAmelCase_ = DDIMScheduler() lowerCAmelCase_ = self.dummy_vqvae_and_unet lowerCAmelCase_ = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_UpperCamelCase , scheduler=_UpperCamelCase ) lowerCAmelCase_ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) np.random.seed(0 ) lowerCAmelCase_ = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) lowerCAmelCase_ = torch.Generator(device=_UpperCamelCase ).manual_seed(42 ) lowerCAmelCase_ = pipe(raw_audio=_UpperCamelCase , generator=_UpperCamelCase , start_step=5 , steps=10 ) lowerCAmelCase_ = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) lowerCAmelCase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowerCAmelCase_ = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 lowerCAmelCase_ = self.dummy_unet_condition lowerCAmelCase_ = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_UpperCamelCase , mel=_UpperCamelCase , scheduler=_UpperCamelCase ) lowerCAmelCase_ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) np.random.seed(0 ) lowerCAmelCase_ = torch.rand((1, 1, 10) ) lowerCAmelCase_ = pipe(generator=_UpperCamelCase , encoding=_UpperCamelCase ) lowerCAmelCase_ = output.images[0] lowerCAmelCase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowerCAmelCase_ = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): def __a ( self ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Optional[int]: lowerCAmelCase_ = torch_device lowerCAmelCase_ = DiffusionPipeline.from_pretrained("teticio/audio-diffusion-ddim-256" ) lowerCAmelCase_ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase_ = torch.Generator(device=_UpperCamelCase ).manual_seed(42 ) lowerCAmelCase_ = pipe(generator=_UpperCamelCase ) lowerCAmelCase_ = output.audios[0] lowerCAmelCase_ = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] lowerCAmelCase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowerCAmelCase_ = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
231
import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, load_metric from torch.utils.data import DataLoader from utils_qa import postprocess_qa_predictions import transformers from transformers import AutoTokenizer, EvalPrediction, default_data_collator, set_seed from transformers.trainer_pt_utils import nested_concat, nested_truncate __UpperCAmelCase : Optional[int] = trt.Logger(trt.Logger.WARNING) __UpperCAmelCase : Tuple = absl_logging.get_absl_logger() absl_logger.setLevel(logging.WARNING) __UpperCAmelCase : Optional[Any] = logging.getLogger(__name__) __UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--onnx_model_path", default=None, type=str, required=True, help="Path to ONNX model: ", ) parser.add_argument( "--output_dir", default=None, type=str, required=True, help="The output directory where the model checkpoints and predictions will be written.", ) # Other parameters parser.add_argument( "--tokenizer_name", default="", type=str, required=True, help="Pretrained tokenizer name or path if not the same as model_name", ) parser.add_argument( "--version_2_with_negative", action="store_true", help="If true, the SQuAD examples contain some that do not have an answer.", ) parser.add_argument( "--null_score_diff_threshold", type=float, default=0.0, help="If null_score - best_non_null is greater than the threshold predict null.", ) parser.add_argument( "--max_seq_length", default=384, type=int, help=( "The maximum total input sequence length after WordPiece tokenization. Sequences " "longer than this will be truncated, and sequences shorter than this will be padded." ), ) parser.add_argument( "--doc_stride", default=128, type=int, help="When splitting up a long document into chunks, how much stride to take between chunks.", ) parser.add_argument("--per_device_eval_batch_size", default=8, type=int, help="Batch size per GPU/CPU for evaluation.") parser.add_argument( "--n_best_size", default=20, type=int, help="The total number of n-best predictions to generate in the nbest_predictions.json output file.", ) parser.add_argument( "--max_answer_length", default=30, type=int, help=( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ), ) parser.add_argument("--seed", type=int, default=42, help="random seed for initialization") parser.add_argument( "--dataset_name", type=str, default=None, required=True, help="The name of the dataset to use (via the datasets library).", ) parser.add_argument( "--dataset_config_name", type=str, default=None, help="The configuration name of the dataset to use (via the datasets library).", ) parser.add_argument( "--preprocessing_num_workers", type=int, default=4, help="A csv or a json file containing the training data." ) parser.add_argument("--overwrite_cache", action="store_true", help="Overwrite the cached training and evaluation sets") parser.add_argument( "--fp16", action="store_true", help="Whether to use 16-bit (mixed) precision instead of 32-bit", ) parser.add_argument( "--int8", action="store_true", help="Whether to use INT8", ) __UpperCAmelCase : Tuple = parser.parse_args() if args.tokenizer_name: __UpperCAmelCase : Tuple = AutoTokenizer.from_pretrained(args.tokenizer_name, use_fast=True) else: raise ValueError( "You are instantiating a new tokenizer from scratch. This is not supported by this script." "You can do it from another script, save it, and load it from here, using --tokenizer_name." ) logger.info("Training/evaluation parameters %s", args) __UpperCAmelCase : Optional[Any] = args.per_device_eval_batch_size __UpperCAmelCase : Dict = (args.eval_batch_size, args.max_seq_length) # TRT Engine properties __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : str = "temp_engine/bert-fp32.engine" if args.fpaa: __UpperCAmelCase : Tuple = "temp_engine/bert-fp16.engine" if args.inta: __UpperCAmelCase : List[Any] = "temp_engine/bert-int8.engine" # import ONNX file if not os.path.exists("temp_engine"): os.makedirs("temp_engine") __UpperCAmelCase : Optional[int] = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH) with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser( network, TRT_LOGGER ) as parser: with open(args.onnx_model_path, "rb") as model: if not parser.parse(model.read()): for error in range(parser.num_errors): print(parser.get_error(error)) # Query input names and shapes from parsed TensorRT network __UpperCAmelCase : int = [network.get_input(i) for i in range(network.num_inputs)] __UpperCAmelCase : List[Any] = [_input.name for _input in network_inputs] # ex: ["actual_input1"] with builder.create_builder_config() as config: __UpperCAmelCase : Optional[Any] = 1 << 50 if STRICT_TYPES: config.set_flag(trt.BuilderFlag.STRICT_TYPES) if args.fpaa: config.set_flag(trt.BuilderFlag.FPaa) if args.inta: config.set_flag(trt.BuilderFlag.INTa) __UpperCAmelCase : Any = builder.create_optimization_profile() config.add_optimization_profile(profile) for i in range(len(input_names)): profile.set_shape(input_names[i], INPUT_SHAPE, INPUT_SHAPE, INPUT_SHAPE) __UpperCAmelCase : Union[str, Any] = builder.build_engine(network, config) # serialize_engine and store in file (can be directly loaded and deserialized): with open(engine_name, "wb") as f: f.write(engine.serialize()) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) -> str: __snake_case: Tuple = np.asarray(inputs["""input_ids"""] , dtype=np.intaa) __snake_case: Union[str, Any] = np.asarray(inputs["""attention_mask"""] , dtype=np.intaa) __snake_case: List[str] = np.asarray(inputs["""token_type_ids"""] , dtype=np.intaa) # Copy inputs cuda.memcpy_htod_async(d_inputs[0] , input_ids.ravel() , SCREAMING_SNAKE_CASE__) cuda.memcpy_htod_async(d_inputs[1] , attention_mask.ravel() , SCREAMING_SNAKE_CASE__) cuda.memcpy_htod_async(d_inputs[2] , token_type_ids.ravel() , SCREAMING_SNAKE_CASE__) # start time __snake_case: int = time.time() # Run inference context.execute_async( bindings=[int(SCREAMING_SNAKE_CASE__) for d_inp in d_inputs] + [int(SCREAMING_SNAKE_CASE__), int(SCREAMING_SNAKE_CASE__)] , stream_handle=stream.handle) # Transfer predictions back from GPU cuda.memcpy_dtoh_async(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) cuda.memcpy_dtoh_async(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) # Synchronize the stream and take time stream.synchronize() # end time __snake_case: Optional[Any] = time.time() __snake_case: Dict = end_time - start_time __snake_case: Any = (h_outputa, h_outputa) # print(outputs) return outputs, infer_time # Initialize the accelerator. We will let the accelerator handle device placement for us in this example. __UpperCAmelCase : Union[str, Any] = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO, ) # Setup logging, we only want one process per machine to log things on the screen. # accelerator.is_local_main_process is only True for one process per machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). if args.dataset_name is not None: # Downloading and loading a dataset from the hub. __UpperCAmelCase : Union[str, Any] = load_dataset(args.dataset_name, args.dataset_config_name) else: raise ValueError("Evaluation requires a dataset name") # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Preprocessing the datasets. # Preprocessing is slighlty different for training and evaluation. __UpperCAmelCase : str = raw_datasets["validation"].column_names __UpperCAmelCase : Dict = "question" if "question" in column_names else column_names[0] __UpperCAmelCase : List[Any] = "context" if "context" in column_names else column_names[1] __UpperCAmelCase : List[str] = "answers" if "answers" in column_names else column_names[2] # Padding side determines if we do (question|context) or (context|question). __UpperCAmelCase : List[str] = tokenizer.padding_side == "right" if args.max_seq_length > tokenizer.model_max_length: logger.warning( f'The max_seq_length passed ({args.max_seq_length}) is larger than the maximum length for the' f'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) __UpperCAmelCase : Union[str, Any] = min(args.max_seq_length, tokenizer.model_max_length) def A__ ( SCREAMING_SNAKE_CASE__) -> Optional[int]: # Some of the questions have lots of whitespace on the left, which is not useful and will make the # truncation of the context fail (the tokenized question will take a lots of space). So we remove that # left whitespace __snake_case: Optional[int] = [q.lstrip() for q in examples[question_column_name]] # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results # in one example possible giving several features when a context is long, each of those features having a # context that overlaps a bit the context of the previous feature. __snake_case: List[str] = tokenizer( examples[question_column_name if pad_on_right else context_column_name] , examples[context_column_name if pad_on_right else question_column_name] , truncation="""only_second""" if pad_on_right else """only_first""" , max_length=SCREAMING_SNAKE_CASE__ , stride=args.doc_stride , return_overflowing_tokens=SCREAMING_SNAKE_CASE__ , return_offsets_mapping=SCREAMING_SNAKE_CASE__ , padding="""max_length""" , ) # Since one example might give us several features if it has a long context, we need a map from a feature to # its corresponding example. This key gives us just that. __snake_case: Optional[Any] = tokenized_examples.pop("""overflow_to_sample_mapping""") # For evaluation, we will need to convert our predictions to substrings of the context, so we keep the # corresponding example_id and we will store the offset mappings. __snake_case: int = [] for i in range(len(tokenized_examples["""input_ids"""])): # Grab the sequence corresponding to that example (to know what is the context and what is the question). __snake_case: int = tokenized_examples.sequence_ids(SCREAMING_SNAKE_CASE__) __snake_case: List[Any] = 1 if pad_on_right else 0 # One example can give several spans, this is the index of the example containing this span of text. __snake_case: Any = sample_mapping[i] tokenized_examples["example_id"].append(examples["""id"""][sample_index]) # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token # position is part of the context or not. __snake_case: Dict = [ (o if sequence_ids[k] == context_index else None) for k, o in enumerate(tokenized_examples["""offset_mapping"""][i]) ] return tokenized_examples __UpperCAmelCase : int = raw_datasets["validation"] # Validation Feature Creation __UpperCAmelCase : Dict = eval_examples.map( prepare_validation_features, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc="Running tokenizer on validation dataset", ) __UpperCAmelCase : Dict = default_data_collator __UpperCAmelCase : List[Any] = eval_dataset.remove_columns(["example_id", "offset_mapping"]) __UpperCAmelCase : str = DataLoader( eval_dataset_for_model, collate_fn=data_collator, batch_size=args.per_device_eval_batch_size ) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__="eval") -> Optional[int]: # Post-processing: we match the start logits and end logits to answers in the original context. __snake_case: Optional[Any] = postprocess_qa_predictions( examples=SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , predictions=SCREAMING_SNAKE_CASE__ , version_2_with_negative=args.version_2_with_negative , n_best_size=args.n_best_size , max_answer_length=args.max_answer_length , null_score_diff_threshold=args.null_score_diff_threshold , output_dir=args.output_dir , prefix=SCREAMING_SNAKE_CASE__ , ) # Format the result to the format the metric expects. if args.version_2_with_negative: __snake_case: Tuple = [ {"""id""": k, """prediction_text""": v, """no_answer_probability""": 0.0} for k, v in predictions.items() ] else: __snake_case: str = [{"""id""": k, """prediction_text""": v} for k, v in predictions.items()] __snake_case: Optional[Any] = [{"""id""": ex["""id"""], """answers""": ex[answer_column_name]} for ex in examples] return EvalPrediction(predictions=SCREAMING_SNAKE_CASE__ , label_ids=SCREAMING_SNAKE_CASE__) __UpperCAmelCase : List[str] = load_metric("squad_v2" if args.version_2_with_negative else "squad") # Evaluation! logger.info("Loading ONNX model %s for evaluation", args.onnx_model_path) with open(engine_name, "rb") as f, trt.Runtime(TRT_LOGGER) as runtime, runtime.deserialize_cuda_engine( f.read() ) as engine, engine.create_execution_context() as context: # setup for TRT inferrence for i in range(len(input_names)): context.set_binding_shape(i, INPUT_SHAPE) assert context.all_binding_shapes_specified def A__ ( SCREAMING_SNAKE_CASE__) -> Union[str, Any]: return trt.volume(engine.get_binding_shape(SCREAMING_SNAKE_CASE__)) * engine.get_binding_dtype(SCREAMING_SNAKE_CASE__).itemsize # Allocate device memory for inputs and outputs. __UpperCAmelCase : int = [cuda.mem_alloc(binding_nbytes(binding)) for binding in engine if engine.binding_is_input(binding)] # Allocate output buffer __UpperCAmelCase : Optional[int] = cuda.pagelocked_empty(tuple(context.get_binding_shape(3)), dtype=np.floataa) __UpperCAmelCase : Any = cuda.pagelocked_empty(tuple(context.get_binding_shape(4)), dtype=np.floataa) __UpperCAmelCase : Union[str, Any] = cuda.mem_alloc(h_outputa.nbytes) __UpperCAmelCase : Optional[Any] = cuda.mem_alloc(h_outputa.nbytes) # Create a stream in which to copy inputs/outputs and run inference. __UpperCAmelCase : Optional[int] = cuda.Stream() # Evaluation logger.info("***** Running Evaluation *****") logger.info(f' Num examples = {len(eval_dataset)}') logger.info(f' Batch size = {args.per_device_eval_batch_size}') __UpperCAmelCase : Optional[Any] = 0.0 __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : Any = timeit.default_timer() __UpperCAmelCase : Union[str, Any] = None for step, batch in enumerate(eval_dataloader): __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = model_infer(batch, context, d_inputs, h_outputa, h_outputa, d_outputa, d_outputa, stream) total_time += infer_time niter += 1 __UpperCAmelCase , __UpperCAmelCase : str = outputs __UpperCAmelCase : Any = torch.tensor(start_logits) __UpperCAmelCase : Tuple = torch.tensor(end_logits) # necessary to pad predictions and labels for being gathered __UpperCAmelCase : Optional[Any] = accelerator.pad_across_processes(start_logits, dim=1, pad_index=-100) __UpperCAmelCase : int = accelerator.pad_across_processes(end_logits, dim=1, pad_index=-100) __UpperCAmelCase : List[str] = (accelerator.gather(start_logits).cpu().numpy(), accelerator.gather(end_logits).cpu().numpy()) __UpperCAmelCase : List[str] = logits if all_preds is None else nested_concat(all_preds, logits, padding_index=-100) if all_preds is not None: __UpperCAmelCase : Union[str, Any] = nested_truncate(all_preds, len(eval_dataset)) __UpperCAmelCase : List[str] = timeit.default_timer() - start_time logger.info(" Evaluation done in total %f secs (%f sec per example)", evalTime, evalTime / len(eval_dataset)) # Inference time from TRT logger.info("Average Inference Time = {:.3f} ms".format(total_time * 1_000 / niter)) logger.info("Total Inference Time = {:.3f} ms".format(total_time * 1_000)) logger.info("Total Number of Inference = %d", niter) __UpperCAmelCase : List[Any] = post_processing_function(eval_examples, eval_dataset, all_preds) __UpperCAmelCase : Optional[int] = metric.compute(predictions=prediction.predictions, references=prediction.label_ids) logger.info(f'Evaluation metrics: {eval_metric}')
111
0
'''simple docstring''' import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin __SCREAMING_SNAKE_CASE :Optional[Any] = logging.get_logger(__name__) enable_full_determinism() class A_ ( _a , _a , unittest.TestCase ): _lowerCamelCase : List[str] = UNetaDModel _lowerCamelCase : Optional[Any] = """sample""" @property def lowercase ( self : int ): _UpperCAmelCase = 4 _UpperCAmelCase = 3 _UpperCAmelCase = (3_2, 3_2) _UpperCAmelCase = floats_tensor((batch_size, num_channels) + sizes ).to(_a ) _UpperCAmelCase = torch.tensor([1_0] ).to(_a ) return {"sample": noise, "timestep": time_step} @property def lowercase ( self : Optional[Any] ): return (3, 3_2, 3_2) @property def lowercase ( self : List[Any] ): return (3, 3_2, 3_2) def lowercase ( self : Tuple ): _UpperCAmelCase = { "block_out_channels": (3_2, 6_4), "down_block_types": ("DownBlock2D", "AttnDownBlock2D"), "up_block_types": ("AttnUpBlock2D", "UpBlock2D"), "attention_head_dim": 3, "out_channels": 3, "in_channels": 3, "layers_per_block": 2, "sample_size": 3_2, } _UpperCAmelCase = self.dummy_input return init_dict, inputs_dict class A_ ( _a , _a , unittest.TestCase ): _lowerCamelCase : Optional[Any] = UNetaDModel _lowerCamelCase : int = """sample""" @property def lowercase ( self : int ): _UpperCAmelCase = 4 _UpperCAmelCase = 4 _UpperCAmelCase = (3_2, 3_2) _UpperCAmelCase = floats_tensor((batch_size, num_channels) + sizes ).to(_a ) _UpperCAmelCase = torch.tensor([1_0] ).to(_a ) return {"sample": noise, "timestep": time_step} @property def lowercase ( self : List[Any] ): return (4, 3_2, 3_2) @property def lowercase ( self : Tuple ): return (4, 3_2, 3_2) def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = { "sample_size": 3_2, "in_channels": 4, "out_channels": 4, "layers_per_block": 2, "block_out_channels": (3_2, 6_4), "attention_head_dim": 3_2, "down_block_types": ("DownBlock2D", "DownBlock2D"), "up_block_types": ("UpBlock2D", "UpBlock2D"), } _UpperCAmelCase = self.dummy_input return init_dict, inputs_dict def lowercase ( self : Optional[Any] ): _UpperCAmelCase = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=_a ) self.assertIsNotNone(_a ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(_a ) _UpperCAmelCase = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != "cuda" , "This test is supposed to run on GPU" ) def lowercase ( self : int ): _UpperCAmelCase = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=_a ) model.to(_a ) _UpperCAmelCase = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != "cuda" , "This test is supposed to run on GPU" ) def lowercase ( self : Union[str, Any] ): # by defautl model loading will use accelerate as `low_cpu_mem_usage=True` _UpperCAmelCase = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=_a ) model_accelerate.to(_a ) model_accelerate.eval() _UpperCAmelCase = torch.randn( 1 , model_accelerate.config.in_channels , model_accelerate.config.sample_size , model_accelerate.config.sample_size , generator=torch.manual_seed(0 ) , ) _UpperCAmelCase = noise.to(_a ) _UpperCAmelCase = torch.tensor([1_0] * noise.shape[0] ).to(_a ) _UpperCAmelCase = model_accelerate(_a , _a )["sample"] # two models don't need to stay in the device at the same time del model_accelerate torch.cuda.empty_cache() gc.collect() _UpperCAmelCase = UNetaDModel.from_pretrained( "fusing/unet-ldm-dummy-update" , output_loading_info=_a , low_cpu_mem_usage=_a ) model_normal_load.to(_a ) model_normal_load.eval() _UpperCAmelCase = model_normal_load(_a , _a )["sample"] assert torch_all_close(_a , _a , rtol=1e-3 ) def lowercase ( self : Optional[int] ): _UpperCAmelCase = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" ) model.eval() model.to(_a ) _UpperCAmelCase = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) _UpperCAmelCase = noise.to(_a ) _UpperCAmelCase = torch.tensor([1_0] * noise.shape[0] ).to(_a ) with torch.no_grad(): _UpperCAmelCase = model(_a , _a ).sample _UpperCAmelCase = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off _UpperCAmelCase = torch.tensor([-1_3.3_2_5_8, -2_0.1_1_0_0, -1_5.9_8_7_3, -1_7.6_6_1_7, -2_3.0_5_9_6, -1_7.9_4_1_9, -1_3.3_6_7_5, -1_6.1_8_8_9, -1_2.3_8_0_0] ) # fmt: on self.assertTrue(torch_all_close(_a , _a , rtol=1e-3 ) ) class A_ ( _a , _a , unittest.TestCase ): _lowerCamelCase : str = UNetaDModel _lowerCamelCase : Dict = """sample""" @property def lowercase ( self : Optional[int] , snake_case_ : List[str]=(3_2, 3_2) ): _UpperCAmelCase = 4 _UpperCAmelCase = 3 _UpperCAmelCase = floats_tensor((batch_size, num_channels) + sizes ).to(_a ) _UpperCAmelCase = torch.tensor(batch_size * [1_0] ).to(dtype=torch.intaa , device=_a ) return {"sample": noise, "timestep": time_step} @property def lowercase ( self : Optional[int] ): return (3, 3_2, 3_2) @property def lowercase ( self : Any ): return (3, 3_2, 3_2) def lowercase ( self : List[Any] ): _UpperCAmelCase = { "block_out_channels": [3_2, 6_4, 6_4, 6_4], "in_channels": 3, "layers_per_block": 1, "out_channels": 3, "time_embedding_type": "fourier", "norm_eps": 1e-6, "mid_block_scale_factor": math.sqrt(2.0 ), "norm_num_groups": None, "down_block_types": [ "SkipDownBlock2D", "AttnSkipDownBlock2D", "SkipDownBlock2D", "SkipDownBlock2D", ], "up_block_types": [ "SkipUpBlock2D", "SkipUpBlock2D", "AttnSkipUpBlock2D", "SkipUpBlock2D", ], } _UpperCAmelCase = self.dummy_input return init_dict, inputs_dict @slow def lowercase ( self : List[Any] ): _UpperCAmelCase = UNetaDModel.from_pretrained("google/ncsnpp-celebahq-256" , output_loading_info=_a ) self.assertIsNotNone(_a ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(_a ) _UpperCAmelCase = self.dummy_input _UpperCAmelCase = floats_tensor((4, 3) + (2_5_6, 2_5_6) ).to(_a ) _UpperCAmelCase = noise _UpperCAmelCase = model(**_a ) assert image is not None, "Make sure output is not None" @slow def lowercase ( self : Dict ): _UpperCAmelCase = UNetaDModel.from_pretrained("google/ncsnpp-celebahq-256" ) model.to(_a ) _UpperCAmelCase = 4 _UpperCAmelCase = 3 _UpperCAmelCase = (2_5_6, 2_5_6) _UpperCAmelCase = torch.ones((batch_size, num_channels) + sizes ).to(_a ) _UpperCAmelCase = torch.tensor(batch_size * [1e-4] ).to(_a ) with torch.no_grad(): _UpperCAmelCase = model(_a , _a ).sample _UpperCAmelCase = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off _UpperCAmelCase = torch.tensor([-4_8_4_2.8_6_9_1, -6_4_9_9.6_6_3_1, -3_8_0_0.1_9_5_3, -7_9_7_8.2_6_8_6, -1_0_9_8_0.7_1_2_9, -2_0_0_2_8.8_5_3_5, 8_1_4_8.2_8_2_2, 2_3_4_2.2_9_0_5, 5_6_7.7_6_0_8] ) # fmt: on self.assertTrue(torch_all_close(_a , _a , rtol=1e-2 ) ) def lowercase ( self : str ): _UpperCAmelCase = UNetaDModel.from_pretrained("fusing/ncsnpp-ffhq-ve-dummy-update" ) model.to(_a ) _UpperCAmelCase = 4 _UpperCAmelCase = 3 _UpperCAmelCase = (3_2, 3_2) _UpperCAmelCase = torch.ones((batch_size, num_channels) + sizes ).to(_a ) _UpperCAmelCase = torch.tensor(batch_size * [1e-4] ).to(_a ) with torch.no_grad(): _UpperCAmelCase = model(_a , _a ).sample _UpperCAmelCase = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off _UpperCAmelCase = torch.tensor([-0.0_3_2_5, -0.0_9_0_0, -0.0_8_6_9, -0.0_3_3_2, -0.0_7_2_5, -0.0_2_7_0, -0.0_1_0_1, 0.0_2_2_7, 0.0_2_5_6] ) # fmt: on self.assertTrue(torch_all_close(_a , _a , rtol=1e-2 ) ) def lowercase ( self : Any ): # not required for this model pass
370
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging __SCREAMING_SNAKE_CASE :Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE :Dict = '''▁''' __SCREAMING_SNAKE_CASE :List[str] = { '''vocab_file''': '''vocab.json''', '''spm_file''': '''sentencepiece.bpe.model''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE :Tuple = { '''vocab_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json''', }, '''spm_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model''', }, '''tokenizer_config_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json''', }, } __SCREAMING_SNAKE_CASE :Optional[int] = { '''facebook/m2m100_418M''': 1024, } # fmt: off __SCREAMING_SNAKE_CASE :Dict = { '''m2m100''': ['''af''', '''am''', '''ar''', '''ast''', '''az''', '''ba''', '''be''', '''bg''', '''bn''', '''br''', '''bs''', '''ca''', '''ceb''', '''cs''', '''cy''', '''da''', '''de''', '''el''', '''en''', '''es''', '''et''', '''fa''', '''ff''', '''fi''', '''fr''', '''fy''', '''ga''', '''gd''', '''gl''', '''gu''', '''ha''', '''he''', '''hi''', '''hr''', '''ht''', '''hu''', '''hy''', '''id''', '''ig''', '''ilo''', '''is''', '''it''', '''ja''', '''jv''', '''ka''', '''kk''', '''km''', '''kn''', '''ko''', '''lb''', '''lg''', '''ln''', '''lo''', '''lt''', '''lv''', '''mg''', '''mk''', '''ml''', '''mn''', '''mr''', '''ms''', '''my''', '''ne''', '''nl''', '''no''', '''ns''', '''oc''', '''or''', '''pa''', '''pl''', '''ps''', '''pt''', '''ro''', '''ru''', '''sd''', '''si''', '''sk''', '''sl''', '''so''', '''sq''', '''sr''', '''ss''', '''su''', '''sv''', '''sw''', '''ta''', '''th''', '''tl''', '''tn''', '''tr''', '''uk''', '''ur''', '''uz''', '''vi''', '''wo''', '''xh''', '''yi''', '''yo''', '''zh''', '''zu'''], '''wmt21''': ['''en''', '''ha''', '''is''', '''ja''', '''cs''', '''ru''', '''zh''', '''de'''] } class A_ ( lowerCAmelCase_ ): _lowerCamelCase : List[str] = VOCAB_FILES_NAMES _lowerCamelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase : List[Any] = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase : List[str] = ["""input_ids""", """attention_mask"""] _lowerCamelCase : List[int] = [] _lowerCamelCase : List[int] = [] def __init__( self : List[str] , snake_case_ : Any , snake_case_ : List[Any] , snake_case_ : str=None , snake_case_ : int=None , snake_case_ : str="<s>" , snake_case_ : int="</s>" , snake_case_ : Any="</s>" , snake_case_ : List[str]="<pad>" , snake_case_ : Optional[int]="<unk>" , snake_case_ : Union[str, Any]="m2m100" , snake_case_ : Optional[Dict[str, Any]] = None , snake_case_ : List[str]=8 , **snake_case_ : str , ): _UpperCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs _UpperCAmelCase = language_codes _UpperCAmelCase = FAIRSEQ_LANGUAGE_CODES[language_codes] _UpperCAmelCase = {lang_code: f'__{lang_code}__' for lang_code in fairseq_language_code} _UpperCAmelCase = kwargs.get("additional_special_tokens" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(snake_case_ ) for lang_code in fairseq_language_code if self.get_lang_token(snake_case_ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=snake_case_ , tgt_lang=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , sep_token=snake_case_ , unk_token=snake_case_ , pad_token=snake_case_ , language_codes=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=snake_case_ , **snake_case_ , ) _UpperCAmelCase = vocab_file _UpperCAmelCase = load_json(snake_case_ ) _UpperCAmelCase = {v: k for k, v in self.encoder.items()} _UpperCAmelCase = spm_file _UpperCAmelCase = load_spm(snake_case_ , self.sp_model_kwargs ) _UpperCAmelCase = len(self.encoder ) _UpperCAmelCase = { self.get_lang_token(snake_case_ ): self.encoder_size + i for i, lang_code in enumerate(snake_case_ ) } _UpperCAmelCase = {lang_code: self.encoder_size + i for i, lang_code in enumerate(snake_case_ )} _UpperCAmelCase = {v: k for k, v in self.lang_token_to_id.items()} _UpperCAmelCase = src_lang if src_lang is not None else "en" _UpperCAmelCase = tgt_lang _UpperCAmelCase = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) _UpperCAmelCase = num_madeup_words @property def lowercase ( self : int ): return len(self.encoder ) + len(self.lang_token_to_id ) @property def lowercase ( self : List[Any] ): return self._src_lang @src_lang.setter def lowercase ( self : str , snake_case_ : str ): _UpperCAmelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase ( self : str , snake_case_ : str ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def lowercase ( self : Optional[Any] , snake_case_ : int ): if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(snake_case_ , self.encoder[self.unk_token] ) def lowercase ( self : Any , snake_case_ : int ): if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(snake_case_ , self.unk_token ) def lowercase ( self : List[str] , snake_case_ : List[str] ): _UpperCAmelCase = [] _UpperCAmelCase = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token _UpperCAmelCase = [] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def lowercase ( self : str , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None , snake_case_ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) _UpperCAmelCase = [1] * len(self.prefix_tokens ) _UpperCAmelCase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(snake_case_ )) + suffix_ones return prefix_ones + ([0] * len(snake_case_ )) + ([0] * len(snake_case_ )) + suffix_ones def lowercase ( self : Optional[int] , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase ( self : Dict ): _UpperCAmelCase = {self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Any ): _UpperCAmelCase = self.__dict__.copy() _UpperCAmelCase = None return state def __setstate__( self : List[str] , snake_case_ : Dict ): _UpperCAmelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): _UpperCAmelCase = {} _UpperCAmelCase = load_spm(self.spm_file , self.sp_model_kwargs ) def lowercase ( self : int , snake_case_ : str , snake_case_ : Optional[str] = None ): _UpperCAmelCase = Path(snake_case_ ) if not save_dir.is_dir(): raise OSError(f'{save_directory} should be a directory' ) _UpperCAmelCase = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) _UpperCAmelCase = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , snake_case_ ) if os.path.abspath(self.spm_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , snake_case_ ) elif not os.path.isfile(self.spm_file ): with open(snake_case_ , "wb" ) as fi: _UpperCAmelCase = self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (str(snake_case_ ), str(snake_case_ )) def lowercase ( self : Dict , snake_case_ : List[str] , snake_case_ : str = "en" , snake_case_ : Optional[List[str]] = None , snake_case_ : str = "ro" , **snake_case_ : Any , ): _UpperCAmelCase = src_lang _UpperCAmelCase = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(snake_case_ , snake_case_ , **snake_case_ ) def lowercase ( self : Tuple , snake_case_ : Optional[Any] , snake_case_ : Optional[str] , snake_case_ : Optional[str] , **snake_case_ : Any ): if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) _UpperCAmelCase = src_lang _UpperCAmelCase = self(snake_case_ , add_special_tokens=snake_case_ , **snake_case_ ) _UpperCAmelCase = self.get_lang_id(snake_case_ ) _UpperCAmelCase = tgt_lang_id return inputs def lowercase ( self : List[str] ): self.set_src_lang_special_tokens(self.src_lang ) def lowercase ( self : Optional[Any] ): self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase ( self : Any , snake_case_ : str ): _UpperCAmelCase = self.get_lang_token(snake_case_ ) _UpperCAmelCase = self.lang_token_to_id[lang_token] _UpperCAmelCase = [self.cur_lang_id] _UpperCAmelCase = [self.eos_token_id] def lowercase ( self : List[Any] , snake_case_ : str ): _UpperCAmelCase = self.get_lang_token(snake_case_ ) _UpperCAmelCase = self.lang_token_to_id[lang_token] _UpperCAmelCase = [self.cur_lang_id] _UpperCAmelCase = [self.eos_token_id] def lowercase ( self : Tuple , snake_case_ : str ): return self.lang_code_to_token[lang] def lowercase ( self : List[str] , snake_case_ : str ): _UpperCAmelCase = self.get_lang_token(snake_case_ ) return self.lang_token_to_id[lang_token] def UpperCAmelCase_ ( __lowercase : str , __lowercase : Dict[str, Any] ) -> sentencepiece.SentencePieceProcessor: '''simple docstring''' _UpperCAmelCase = sentencepiece.SentencePieceProcessor(**__lowercase ) spm.Load(str(__lowercase ) ) return spm def UpperCAmelCase_ ( __lowercase : str ) -> Union[Dict, List]: '''simple docstring''' with open(__lowercase , "r" ) as f: return json.load(__lowercase ) def UpperCAmelCase_ ( __lowercase : str , __lowercase : str ) -> None: '''simple docstring''' with open(__lowercase , "w" ) as f: json.dump(__lowercase , __lowercase , indent=2 )
156
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase :int = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase :int = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys lowerCamelCase :Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
206
'''simple docstring''' import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class lowerCamelCase_ ( unittest.TestCase ): def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() UpperCAmelCase__ : str = dict(zip(_A , range(len(_A ) ) ) ) UpperCAmelCase__ : int = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } UpperCAmelCase__ : List[Any] = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 16_000, '''return_attention_mask''': False, '''do_normalize''': True, } UpperCAmelCase__ : str = tempfile.mkdtemp() UpperCAmelCase__ : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) UpperCAmelCase__ : List[Any] = os.path.join(self.tmpdirname , _A ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_A ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_A ) + '''\n''' ) # load decoder from hub UpperCAmelCase__ : Tuple = '''hf-internal-testing/ngram-beam-search-decoder''' def lowercase_ ( self : Union[str, Any] , **_A : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.add_kwargs_tokens_map.copy() kwargs.update(_A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **_A ) def lowercase_ ( self : Dict , **_A : str ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **_A ) def lowercase_ ( self : Tuple , **_A : int ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **_A ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Dict = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , _A ) def lowercase_ ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def lowercase_ ( self : str ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(_A , '''include''' ): WavaVecaProcessorWithLM( tokenizer=_A , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.get_feature_extractor() UpperCAmelCase__ : List[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : str = feature_extractor(_A , return_tensors='''np''' ) UpperCAmelCase__ : Optional[Any] = processor(_A , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.get_feature_extractor() UpperCAmelCase__ : Any = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : Union[str, Any] = '''This is a test string''' UpperCAmelCase__ : Union[str, Any] = processor(text=_A ) UpperCAmelCase__ : Dict = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowercase_ ( self : str , _A : int=(2, 10, 16) , _A : Optional[int]=77 ): '''simple docstring''' np.random.seed(_A ) return np.random.rand(*_A ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Dict = self.get_tokenizer() UpperCAmelCase__ : Tuple = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits(shape=(10, 16) , seed=13 ) UpperCAmelCase__ : Union[str, Any] = processor.decode(_A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(_A )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def lowercase_ ( self : Dict , _A : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(_A ) else: with get_context(_A ).Pool() as pool: UpperCAmelCase__ : List[Any] = processor.batch_decode(_A , _A ) UpperCAmelCase__ : List[str] = list(_A ) with get_context('''fork''' ).Pool() as p: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch(_A , _A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(_A , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(_A , decoded_processor.logit_score ) self.assertListEqual(_A , decoded_processor.lm_score ) def lowercase_ ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : List[Any] = self._get_dummy_logits() UpperCAmelCase__ : Optional[Any] = 15 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : List[str] = -4.0 UpperCAmelCase__ : str = processor.batch_decode( _A , beam_width=_A , beam_prune_logp=_A , token_min_logp=_A , ) UpperCAmelCase__ : Union[str, Any] = decoded_processor_out.text UpperCAmelCase__ : Tuple = list(_A ) with get_context('''fork''' ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( _A , _A , beam_width=_A , beam_prune_logp=_A , token_min_logp=_A , ) UpperCAmelCase__ : Optional[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : Tuple = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(_A , _A ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , _A ) self.assertTrue(np.array_equal(_A , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] , _A , atol=1e-3 ) ) self.assertTrue(np.array_equal(_A , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] , _A , atol=1e-3 ) ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_tokenizer() UpperCAmelCase__ : List[str] = self.get_decoder() UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() UpperCAmelCase__ : List[Any] = 2.0 UpperCAmelCase__ : Union[str, Any] = 5.0 UpperCAmelCase__ : Any = -2_0.0 UpperCAmelCase__ : List[Any] = True UpperCAmelCase__ : Any = processor.batch_decode( _A , alpha=_A , beta=_A , unk_score_offset=_A , lm_score_boundary=_A , ) UpperCAmelCase__ : List[Any] = decoded_processor_out.text UpperCAmelCase__ : List[Any] = list(_A ) decoder.reset_params( alpha=_A , beta=_A , unk_score_offset=_A , lm_score_boundary=_A , ) with get_context('''fork''' ).Pool() as pool: UpperCAmelCase__ : Any = decoder.decode_beams_batch( _A , _A , ) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(_A , _A ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , _A ) UpperCAmelCase__ : Any = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -2_0.0 ) self.assertEqual(lm_model.score_boundary , _A ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : List[Any] = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[Any] = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() UpperCAmelCase__ : Dict = os.listdir(_A ) UpperCAmelCase__ : str = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(_A , _A ) def lowercase_ ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : str = snapshot_download('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained(_A ) UpperCAmelCase__ : Optional[Any] = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Union[str, Any] = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[Any] = os.listdir(_A ) UpperCAmelCase__ : int = os.listdir(_A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(_A , _A ) def lowercase_ ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Any = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Tuple = floats_list((3, 1_000) ) UpperCAmelCase__ : Tuple = processor_wavaveca(_A , return_tensors='''np''' ) UpperCAmelCase__ : Union[str, Any] = processor_auto(_A , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2 ) UpperCAmelCase__ : str = self._get_dummy_logits() UpperCAmelCase__ : List[Any] = processor_wavaveca.batch_decode(_A ) UpperCAmelCase__ : List[Any] = processor_auto.batch_decode(_A ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def lowercase_ ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Any = self.get_tokenizer() UpperCAmelCase__ : List[str] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=_A , feature_extractor=_A , decoder=_A ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def lowercase_ ( _A : Tuple , _A : Tuple ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : Tuple = self._get_dummy_logits()[0] UpperCAmelCase__ : int = processor.decode(_A , output_word_offsets=_A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(_A , _A ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def lowercase_ ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Dict = processor.batch_decode(_A , output_word_offsets=_A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(_A , _A ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(_A , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def lowercase_ ( self : str ): '''simple docstring''' import torch UpperCAmelCase__ : Dict = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=_A ) UpperCAmelCase__ : Optional[Any] = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Any = iter(_A ) UpperCAmelCase__ : Dict = next(_A ) UpperCAmelCase__ : Optional[int] = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) UpperCAmelCase__ : List[Any] = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Union[str, Any] = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): UpperCAmelCase__ : Dict = model(_A ).logits.cpu().numpy() UpperCAmelCase__ : List[str] = processor.decode(logits[0] , output_word_offsets=_A ) UpperCAmelCase__ : Tuple = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : List[str] = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] UpperCAmelCase__ : Tuple = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(_A , '''word''' ) ) , _A ) self.assertEqual(''' '''.join(self.get_from_offsets(_A , '''word''' ) ) , output.text ) # output times UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(_A , '''start_time''' ) ) UpperCAmelCase__ : Dict = torch.tensor(self.get_from_offsets(_A , '''end_time''' ) ) # fmt: off UpperCAmelCase__ : Any = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : Any = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(_A , _A , atol=0.0_1 ) ) self.assertTrue(torch.allclose(_A , _A , atol=0.0_1 ) )
181
0
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __lowerCamelCase ( a_ : Union[str, Any] , a_ : Tuple=None ) -> Any: __SCREAMING_SNAKE_CASE :int = None if token is not None: __SCREAMING_SNAKE_CASE :Any = {'''Accept''': '''application/vnd.github+json''', '''Authorization''': f'''Bearer {token}'''} __SCREAMING_SNAKE_CASE :Any = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100''' __SCREAMING_SNAKE_CASE :Any = requests.get(a_ , headers=a_ ).json() __SCREAMING_SNAKE_CASE :Optional[Any] = {} try: job_links.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) __SCREAMING_SNAKE_CASE :Any = math.ceil((result['''total_count'''] - 1_00) / 1_00 ) for i in range(a_ ): __SCREAMING_SNAKE_CASE :Tuple = requests.get(url + f'''&page={i + 2}''' , headers=a_ ).json() job_links.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) return job_links except Exception: print(f'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} def __lowerCamelCase ( a_ : Dict , a_ : Optional[Any]=None ) -> Optional[int]: __SCREAMING_SNAKE_CASE :Optional[int] = None if token is not None: __SCREAMING_SNAKE_CASE :Optional[int] = {'''Accept''': '''application/vnd.github+json''', '''Authorization''': f'''Bearer {token}'''} __SCREAMING_SNAKE_CASE :int = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100''' __SCREAMING_SNAKE_CASE :List[str] = requests.get(a_ , headers=a_ ).json() __SCREAMING_SNAKE_CASE :Optional[int] = {} try: artifacts.update({artifact['''name''']: artifact['''archive_download_url'''] for artifact in result['''artifacts''']} ) __SCREAMING_SNAKE_CASE :Optional[int] = math.ceil((result['''total_count'''] - 1_00) / 1_00 ) for i in range(a_ ): __SCREAMING_SNAKE_CASE :Union[str, Any] = requests.get(url + f'''&page={i + 2}''' , headers=a_ ).json() artifacts.update({artifact['''name''']: artifact['''archive_download_url'''] for artifact in result['''artifacts''']} ) return artifacts except Exception: print(f'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} def __lowerCamelCase ( a_ : int , a_ : Dict , a_ : str , a_ : str ) -> str: __SCREAMING_SNAKE_CASE :str = None if token is not None: __SCREAMING_SNAKE_CASE :Any = {'''Accept''': '''application/vnd.github+json''', '''Authorization''': f'''Bearer {token}'''} __SCREAMING_SNAKE_CASE :List[Any] = requests.get(a_ , headers=a_ , allow_redirects=a_ ) __SCREAMING_SNAKE_CASE :int = result.headers['''Location'''] __SCREAMING_SNAKE_CASE :str = requests.get(a_ , allow_redirects=a_ ) __SCREAMING_SNAKE_CASE :Optional[int] = os.path.join(a_ , f'''{artifact_name}.zip''' ) with open(a_ , '''wb''' ) as fp: fp.write(response.content ) def __lowerCamelCase ( a_ : Optional[int] , a_ : Optional[Any]=None ) -> int: __SCREAMING_SNAKE_CASE :List[Any] = [] __SCREAMING_SNAKE_CASE :Optional[Any] = [] __SCREAMING_SNAKE_CASE :str = None with zipfile.ZipFile(a_ ) as z: for filename in z.namelist(): if not os.path.isdir(a_ ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(a_ ) as f: for line in f: __SCREAMING_SNAKE_CASE :Any = line.decode('''UTF-8''' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs __SCREAMING_SNAKE_CASE :Tuple = line[: line.index(''': ''' )] __SCREAMING_SNAKE_CASE :List[Any] = line[line.index(''': ''' ) + len(''': ''' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('''FAILED ''' ): # `test` is the test method that failed __SCREAMING_SNAKE_CASE :int = line[len('''FAILED ''' ) :] failed_tests.append(a_ ) elif filename == "job_name.txt": __SCREAMING_SNAKE_CASE :int = line if len(a_ ) != len(a_ ): raise ValueError( f'''`errors` and `failed_tests` should have the same number of elements. Got {len(a_ )} for `errors` ''' f'''and {len(a_ )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some''' ''' problem.''' ) __SCREAMING_SNAKE_CASE :int = None if job_name and job_links: __SCREAMING_SNAKE_CASE :Optional[int] = job_links.get(a_ , a_ ) # A list with elements of the form (line of error, error, failed test) __SCREAMING_SNAKE_CASE :Tuple = [x + [y] + [job_link] for x, y in zip(a_ , a_ )] return result def __lowerCamelCase ( a_ : Tuple , a_ : Any=None ) -> List[str]: __SCREAMING_SNAKE_CASE :List[Any] = [] __SCREAMING_SNAKE_CASE :Dict = [os.path.join(a_ , a_ ) for p in os.listdir(a_ ) if p.endswith('''.zip''' )] for p in paths: errors.extend(get_errors_from_single_artifact(a_ , job_links=a_ ) ) return errors def __lowerCamelCase ( a_ : Optional[Any] , a_ : List[Any]=None ) -> Tuple: __SCREAMING_SNAKE_CASE :Tuple = Counter() counter.update([x[1] for x in logs] ) __SCREAMING_SNAKE_CASE :int = counter.most_common() __SCREAMING_SNAKE_CASE :List[Any] = {} for error, count in counts: if error_filter is None or error not in error_filter: __SCREAMING_SNAKE_CASE :Optional[int] = {'''count''': count, '''failed_tests''': [(x[2], x[0]) for x in logs if x[1] == error]} __SCREAMING_SNAKE_CASE :Optional[Any] = dict(sorted(r.items() , key=lambda a_ : item[1]["count"] , reverse=a_ ) ) return r def __lowerCamelCase ( a_ : Optional[Any] ) -> Tuple: __SCREAMING_SNAKE_CASE :Dict = test.split('''::''' )[0] if test.startswith('''tests/models/''' ): __SCREAMING_SNAKE_CASE :Optional[int] = test.split('''/''' )[2] else: __SCREAMING_SNAKE_CASE :Tuple = None return test def __lowerCamelCase ( a_ : int , a_ : Optional[Any]=None ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE :Optional[Any] = [(x[0], x[1], get_model(x[2] )) for x in logs] __SCREAMING_SNAKE_CASE :Tuple = [x for x in logs if x[2] is not None] __SCREAMING_SNAKE_CASE :List[str] = {x[2] for x in logs} __SCREAMING_SNAKE_CASE :Optional[Any] = {} for test in tests: __SCREAMING_SNAKE_CASE :int = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) __SCREAMING_SNAKE_CASE :Optional[Any] = counter.most_common() __SCREAMING_SNAKE_CASE :Any = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} __SCREAMING_SNAKE_CASE :Dict = sum(error_counts.values() ) if n_errors > 0: __SCREAMING_SNAKE_CASE :Optional[int] = {'''count''': n_errors, '''errors''': error_counts} __SCREAMING_SNAKE_CASE :List[str] = dict(sorted(r.items() , key=lambda a_ : item[1]["count"] , reverse=a_ ) ) return r def __lowerCamelCase ( a_ : Any ) -> Optional[int]: __SCREAMING_SNAKE_CASE :Union[str, Any] = '''| no. | error | status |''' __SCREAMING_SNAKE_CASE :Tuple = '''|-:|:-|:-|''' __SCREAMING_SNAKE_CASE :List[str] = [header, sep] for error in reduced_by_error: __SCREAMING_SNAKE_CASE :Optional[Any] = reduced_by_error[error]['''count'''] __SCREAMING_SNAKE_CASE :Dict = f'''| {count} | {error[:1_00]} | |''' lines.append(a_ ) return "\n".join(a_ ) def __lowerCamelCase ( a_ : Tuple ) -> Tuple: __SCREAMING_SNAKE_CASE :List[str] = '''| model | no. of errors | major error | count |''' __SCREAMING_SNAKE_CASE :str = '''|-:|-:|-:|-:|''' __SCREAMING_SNAKE_CASE :List[str] = [header, sep] for model in reduced_by_model: __SCREAMING_SNAKE_CASE :Optional[Any] = reduced_by_model[model]['''count'''] __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :List[Any] = list(reduced_by_model[model]['''errors'''].items() )[0] __SCREAMING_SNAKE_CASE :List[str] = f'''| {model} | {count} | {error[:60]} | {_count} |''' lines.append(a_ ) return "\n".join(a_ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument("--workflow_run_id", type=str, required=True, help="A GitHub Actions workflow run id.") parser.add_argument( "--output_dir", type=str, required=True, help="Where to store the downloaded artifacts and other result files.", ) parser.add_argument("--token", default=None, type=str, help="A token that has actions:read permission.") lowerCamelCase_ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowerCamelCase_ = get_job_links(args.workflow_run_id, token=args.token) lowerCamelCase_ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowerCamelCase_ = k.find(" / ") lowerCamelCase_ = k[index + len(" / ") :] lowerCamelCase_ = v with open(os.path.join(args.output_dir, "job_links.json"), "w", encoding="UTF-8") as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowerCamelCase_ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, "artifacts.json"), "w", encoding="UTF-8") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowerCamelCase_ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowerCamelCase_ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowerCamelCase_ = counter.most_common(3_0) for item in most_common: print(item) with open(os.path.join(args.output_dir, "errors.json"), "w", encoding="UTF-8") as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowerCamelCase_ = reduce_by_error(errors) lowerCamelCase_ = reduce_by_model(errors) lowerCamelCase_ = make_github_table(reduced_by_error) lowerCamelCase_ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, "reduced_by_error.txt"), "w", encoding="UTF-8") as fp: fp.write(sa) with open(os.path.join(args.output_dir, "reduced_by_model.txt"), "w", encoding="UTF-8") as fp: fp.write(sa)
239
"""simple docstring""" import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() lowerCamelCase_ = logging.get_logger(__name__) def __lowerCamelCase ( a_ : Optional[int] , a_ : Optional[int] , a_ : Any ) -> List[Any]: __SCREAMING_SNAKE_CASE :Any = UniSpeechSatForSequenceClassification.from_pretrained(a_ , config=a_ ) __SCREAMING_SNAKE_CASE :int = downstream_dict['''projector.weight'''] __SCREAMING_SNAKE_CASE :List[Any] = downstream_dict['''projector.bias'''] __SCREAMING_SNAKE_CASE :Union[str, Any] = downstream_dict['''model.post_net.linear.weight'''] __SCREAMING_SNAKE_CASE :List[str] = downstream_dict['''model.post_net.linear.bias'''] return model def __lowerCamelCase ( a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ) -> Optional[int]: __SCREAMING_SNAKE_CASE :Any = UniSpeechSatForAudioFrameClassification.from_pretrained(a_ , config=a_ ) __SCREAMING_SNAKE_CASE :List[str] = downstream_dict['''model.linear.weight'''] __SCREAMING_SNAKE_CASE :Union[str, Any] = downstream_dict['''model.linear.bias'''] return model def __lowerCamelCase ( a_ : Optional[int] , a_ : Optional[Any] , a_ : int ) -> List[str]: __SCREAMING_SNAKE_CASE :List[str] = UniSpeechSatForXVector.from_pretrained(a_ , config=a_ ) __SCREAMING_SNAKE_CASE :Optional[int] = downstream_dict['''connector.weight'''] __SCREAMING_SNAKE_CASE :Tuple = downstream_dict['''connector.bias'''] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): __SCREAMING_SNAKE_CASE :str = downstream_dict[ f'''model.framelevel_feature_extractor.module.{i}.kernel.weight''' ] __SCREAMING_SNAKE_CASE :int = downstream_dict[f'''model.framelevel_feature_extractor.module.{i}.kernel.bias'''] __SCREAMING_SNAKE_CASE :Any = downstream_dict['''model.utterancelevel_feature_extractor.linear1.weight'''] __SCREAMING_SNAKE_CASE :Optional[Any] = downstream_dict['''model.utterancelevel_feature_extractor.linear1.bias'''] __SCREAMING_SNAKE_CASE :Dict = downstream_dict['''model.utterancelevel_feature_extractor.linear2.weight'''] __SCREAMING_SNAKE_CASE :Optional[int] = downstream_dict['''model.utterancelevel_feature_extractor.linear2.bias'''] __SCREAMING_SNAKE_CASE :str = downstream_dict['''objective.W'''] return model @torch.no_grad() def __lowerCamelCase ( a_ : Optional[int] , a_ : Union[str, Any] , a_ : Any , a_ : Union[str, Any] ) -> List[str]: __SCREAMING_SNAKE_CASE :str = torch.load(a_ , map_location='''cpu''' ) __SCREAMING_SNAKE_CASE :str = checkpoint['''Downstream'''] __SCREAMING_SNAKE_CASE :str = UniSpeechSatConfig.from_pretrained(a_ ) __SCREAMING_SNAKE_CASE :List[str] = WavaVecaFeatureExtractor.from_pretrained( a_ , return_attention_mask=a_ , do_normalize=a_ ) __SCREAMING_SNAKE_CASE :Optional[Any] = hf_config.architectures[0] if arch.endswith('''ForSequenceClassification''' ): __SCREAMING_SNAKE_CASE :str = convert_classification(a_ , a_ , a_ ) elif arch.endswith('''ForAudioFrameClassification''' ): __SCREAMING_SNAKE_CASE :Tuple = convert_diarization(a_ , a_ , a_ ) elif arch.endswith('''ForXVector''' ): __SCREAMING_SNAKE_CASE :List[Any] = convert_xvector(a_ , a_ , a_ ) else: raise NotImplementedError(f'''S3PRL weights conversion is not supported for {arch}''' ) if hf_config.use_weighted_layer_sum: __SCREAMING_SNAKE_CASE :Dict = checkpoint['''Featurizer''']['''weights'''] hf_feature_extractor.save_pretrained(a_ ) hf_model.save_pretrained(a_ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( "--base_model_name", default=None, type=str, help="Name of the huggingface pretrained base model." ) parser.add_argument("--config_path", default=None, type=str, help="Path to the huggingface classifier config.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to the s3prl checkpoint.") parser.add_argument("--model_dump_path", default=None, type=str, help="Path to the final converted model.") lowerCamelCase_ = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
239
1
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar('''_T''') class SCREAMING_SNAKE_CASE__ ( Generic[_T] ): """simple docstring""" def __init__( self , snake_case__ = None ): """simple docstring""" lowerCAmelCase : list[_T] = list(iterable or [] ) lowerCAmelCase : list[_T] = [] def __len__( self ): """simple docstring""" return len(self._stacka ) + len(self._stacka ) def __repr__( self ): """simple docstring""" return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def lowercase__ ( self , snake_case__ ): """simple docstring""" self._stacka.append(snake_case__ ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = self._stacka.pop lowerCAmelCase : List[str] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("Queue is empty" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
108
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' if principal <= 0: raise Exception("Principal borrowed must be > 0" ) if rate_per_annum < 0: raise Exception("Rate of interest must be >= 0" ) if years_to_repay <= 0 or not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise Exception("Years to repay must be an integer > 0" ) # Yearly rate is divided by 12 to get monthly rate lowerCAmelCase : Tuple = rate_per_annum / 1_2 # Years to repay is multiplied by 12 to get number of payments as payment is monthly lowerCAmelCase : List[Any] = years_to_repay * 1_2 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
108
1
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy SCREAMING_SNAKE_CASE__ = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: torch.nn.Module , __lowerCamelCase: BnbQuantizationConfig , __lowerCamelCase: Union[str, os.PathLike] = None , __lowerCamelCase: Optional[Dict[str, Union[int, str, torch.device]]] = None , __lowerCamelCase: Optional[List[str]] = None , __lowerCamelCase: Optional[Dict[Union[int, str], Union[int, str]]] = None , __lowerCamelCase: Optional[Union[str, os.PathLike]] = None , __lowerCamelCase: bool = False , ): '''simple docstring''' lowercase_ = bnb_quantization_config.load_in_abit lowercase_ = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( "You have a version of `bitsandbytes` that is not compatible with 8bit quantization," " make sure you have the latest version of `bitsandbytes` installed." ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( "You have a version of `bitsandbytes` that is not compatible with 4bit quantization," "make sure you have the latest version of `bitsandbytes` installed." ) lowercase_ = [] # custom device map if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(device_map.keys() ) > 1: lowercase_ = [key for key, value in device_map.items() if value in ["disk", "cpu"]] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: lowercase_ = get_keys_to_not_convert(__lowerCamelCase ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(__lowerCamelCase ) lowercase_ = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: lowercase_ = [] lowercase_ = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(__lowerCamelCase ) # compatibility with peft lowercase_ = load_in_abit lowercase_ = load_in_abit lowercase_ = get_parameter_device(__lowerCamelCase ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( "It is not recommended to quantize a loaded model. " "The model should be instantiated under the `init_empty_weights` context manager." ) lowercase_ = replace_with_bnb_layers(__lowerCamelCase , __lowerCamelCase , modules_to_not_convert=__lowerCamelCase ) # convert param to the right dtype lowercase_ = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: lowercase_ = name.replace(".weight" , "" ).replace(".bias" , "" ) lowercase_ = getattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(__lowerCamelCase ): param.to(__lowerCamelCase ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError("No GPU found. A GPU is needed for quantization." ) logger.info( F'The model device type is {model_device.type}. However, cuda is needed for quantization.' "We move the model to cuda." ) return model elif weights_location is None: raise RuntimeError( F'`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} ' ) else: with init_empty_weights(): lowercase_ = replace_with_bnb_layers( __lowerCamelCase , __lowerCamelCase , modules_to_not_convert=__lowerCamelCase ) lowercase_ = get_quantized_model_device_map( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , max_memory=__lowerCamelCase , no_split_module_classes=__lowerCamelCase , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): lowercase_ = True lowercase_ = any(x in list(device_map.values() ) for x in ["cpu", "disk"] ) load_checkpoint_in_model( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , dtype=bnb_quantization_config.torch_dtype , offload_folder=__lowerCamelCase , offload_state_dict=__lowerCamelCase , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(__lowerCamelCase , device_map=__lowerCamelCase , offload_dir=__lowerCamelCase ) def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: List[Any] , __lowerCamelCase: str , __lowerCamelCase: Union[str, Any]=None , __lowerCamelCase: Optional[Any]=None , __lowerCamelCase: str=None ): '''simple docstring''' if device_map is None: if torch.cuda.is_available(): lowercase_ = {"": torch.cuda.current_device()} else: raise RuntimeError("No GPU found. A GPU is needed for quantization." ) logger.info("The device_map was not initialized." "Setting device_map to `{'':torch.cuda.current_device()}`." ) if isinstance(__lowerCamelCase , __lowerCamelCase ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( "If passing a string for `device_map`, please choose 'auto', 'balanced', 'balanced_low_0' or " "'sequential'." ) lowercase_ = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) lowercase_ = {} lowercase_ = special_dtypes lowercase_ = no_split_module_classes lowercase_ = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": lowercase_ = get_balanced_memory( __lowerCamelCase , low_zero=(device_map == "balanced_low_0") , max_memory=__lowerCamelCase , **__lowerCamelCase , ) lowercase_ = max_memory lowercase_ = infer_auto_device_map(__lowerCamelCase , **__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): # check if don't have any quantized module on the cpu lowercase_ = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules lowercase_ = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( "\n Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit\n the quantized model. If you want to dispatch the model on the CPU or the disk while keeping\n these modules in `torch_dtype`, you need to pass a custom `device_map` to\n `load_and_quantize_model`. Check\n https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk\n for more details.\n " ) else: logger.info( "Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit" ) del device_map_without_some_modules return device_map def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: List[str] , __lowerCamelCase: Any , __lowerCamelCase: Dict=None , __lowerCamelCase: int=None ): '''simple docstring''' if modules_to_not_convert is None: lowercase_ = [] lowercase_ , lowercase_ = _replace_with_bnb_layers( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if not has_been_replaced: logger.warning( "You are loading your model in 8bit or 4bit but no linear modules were found in your model." " this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers." " Please double check your model architecture, or submit an issue on github if you think this is" " a bug." ) return model def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: int , __lowerCamelCase: List[str] , __lowerCamelCase: Union[str, Any]=None , __lowerCamelCase: Dict=None , ): '''simple docstring''' lowercase_ = False for name, module in model.named_children(): if current_key_name is None: lowercase_ = [] current_key_name.append(__lowerCamelCase ) if isinstance(__lowerCamelCase , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` lowercase_ = ".".join(__lowerCamelCase ) lowercase_ = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: lowercase_ = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: lowercase_ = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=__lowerCamelCase , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: lowercase_ = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError("load_in_8bit and load_in_4bit can't be both False" ) lowercase_ = module.weight.data if module.bias is not None: lowercase_ = module.bias.data bnb_module.requires_grad_(__lowerCamelCase ) setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) lowercase_ = True if len(list(module.children() ) ) > 0: lowercase_ , lowercase_ = _replace_with_bnb_layers( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) lowercase_ = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: int ): '''simple docstring''' with init_empty_weights(): lowercase_ = deepcopy(__lowerCamelCase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` lowercase_ = find_tied_parameters(__lowerCamelCase ) # For compatibility with Accelerate < 0.18 if isinstance(__lowerCamelCase , __lowerCamelCase ): lowercase_ = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: lowercase_ = sum(__lowerCamelCase , [] ) lowercase_ = len(__lowerCamelCase ) > 0 # Check if it is a base model lowercase_ = False if hasattr(__lowerCamelCase , "base_model_prefix" ): lowercase_ = not hasattr(__lowerCamelCase , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head lowercase_ = list(model.named_children() ) lowercase_ = [list_modules[-1][0]] # add last module together with tied weights lowercase_ = set(__lowerCamelCase ) - set(__lowerCamelCase ) lowercase_ = list(set(__lowerCamelCase ) ) + list(__lowerCamelCase ) # remove ".weight" from the keys lowercase_ = [".weight", ".bias"] lowercase_ = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: lowercase_ = name.replace(__lowerCamelCase , "" ) filtered_module_names.append(__lowerCamelCase ) return filtered_module_names def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Optional[int] ): '''simple docstring''' for m in model.modules(): if isinstance(__lowerCamelCase , bnb.nn.Linearabit ): return True return False def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: nn.Module ): '''simple docstring''' return next(parameter.parameters() ).device def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: List[str] , __lowerCamelCase: int , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: List[Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: List[str] ): '''simple docstring''' if fpaa_statistics is None: set_module_tensor_to_device(__lowerCamelCase , __lowerCamelCase , 0 , dtype=__lowerCamelCase , value=__lowerCamelCase ) lowercase_ = param_name lowercase_ = model if "." in tensor_name: lowercase_ = tensor_name.split("." ) for split in splits[:-1]: lowercase_ = getattr(__lowerCamelCase , __lowerCamelCase ) if new_module is None: raise ValueError(F'{module} has no attribute {split}.' ) lowercase_ = new_module lowercase_ = splits[-1] # offload weights lowercase_ = False offload_weight(module._parameters[tensor_name] , __lowerCamelCase , __lowerCamelCase , index=__lowerCamelCase ) if hasattr(module._parameters[tensor_name] , "SCB" ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace("weight" , "SCB" ) , __lowerCamelCase , index=__lowerCamelCase , ) else: offload_weight(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index=__lowerCamelCase ) offload_weight(__lowerCamelCase , param_name.replace("weight" , "SCB" ) , __lowerCamelCase , index=__lowerCamelCase ) set_module_tensor_to_device(__lowerCamelCase , __lowerCamelCase , "meta" , dtype=__lowerCamelCase , value=torch.empty(*param.size() ) )
367
import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" def A__ ( self ) -> Any: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def A__ ( self ) -> Tuple: '''simple docstring''' torch.manual_seed(0 ) lowercase_ = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("AttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "AttnUpBlock2D") , ) return model @property def A__ ( self ) -> Tuple: '''simple docstring''' torch.manual_seed(0 ) lowercase_ = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , cross_attention_dim=10 , ) return model @property def A__ ( self ) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0 ) lowercase_ = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("DownEncoderBlock2D", "DownEncoderBlock2D") , up_block_types=("UpDecoderBlock2D", "UpDecoderBlock2D") , ) lowercase_ = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("AttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "AttnUpBlock2D") , ) return vqvae, unet @slow def A__ ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase_ = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) lowercase_ = DDPMScheduler() lowercase_ = AudioDiffusionPipeline(vqvae=UpperCAmelCase , unet=self.dummy_unet , mel=UpperCAmelCase , scheduler=UpperCAmelCase ) lowercase_ = pipe.to(UpperCAmelCase ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) lowercase_ = torch.Generator(device=UpperCAmelCase ).manual_seed(42 ) lowercase_ = pipe(generator=UpperCAmelCase , steps=4 ) lowercase_ = output.audios[0] lowercase_ = output.images[0] lowercase_ = torch.Generator(device=UpperCAmelCase ).manual_seed(42 ) lowercase_ = pipe(generator=UpperCAmelCase , steps=4 , return_dict=UpperCAmelCase ) lowercase_ = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) lowercase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowercase_ = np.frombuffer(image_from_tuple.tobytes() , dtype="uint8" )[:10] lowercase_ = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 lowercase_ = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) lowercase_ = DDIMScheduler() lowercase_ = self.dummy_vqvae_and_unet lowercase_ = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=UpperCAmelCase , scheduler=UpperCAmelCase ) lowercase_ = pipe.to(UpperCAmelCase ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) np.random.seed(0 ) lowercase_ = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) lowercase_ = torch.Generator(device=UpperCAmelCase ).manual_seed(42 ) lowercase_ = pipe(raw_audio=UpperCAmelCase , generator=UpperCAmelCase , start_step=5 , steps=10 ) lowercase_ = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) lowercase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowercase_ = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 lowercase_ = self.dummy_unet_condition lowercase_ = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=UpperCAmelCase , mel=UpperCAmelCase , scheduler=UpperCAmelCase ) lowercase_ = pipe.to(UpperCAmelCase ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) np.random.seed(0 ) lowercase_ = torch.rand((1, 1, 10) ) lowercase_ = pipe(generator=UpperCAmelCase , encoding=UpperCAmelCase ) lowercase_ = output.images[0] lowercase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowercase_ = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" def A__ ( self ) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self ) -> Dict: '''simple docstring''' lowercase_ = torch_device lowercase_ = DiffusionPipeline.from_pretrained("teticio/audio-diffusion-ddim-256" ) lowercase_ = pipe.to(UpperCAmelCase ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) lowercase_ = torch.Generator(device=UpperCAmelCase ).manual_seed(42 ) lowercase_ = pipe(generator=UpperCAmelCase ) lowercase_ = output.audios[0] lowercase_ = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] lowercase_ = np.frombuffer(image.tobytes() , dtype="uint8" )[:10] lowercase_ = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
297
0
'''simple docstring''' import argparse import collections import json import os import re import string import sys import numpy as np lowerCamelCase : List[Any] = re.compile(r"\b(a|an|the)\b", re.UNICODE) lowerCamelCase : Dict = None def _lowerCAmelCase ( ) -> List[str]: """simple docstring""" _SCREAMING_SNAKE_CASE =argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_UpperCamelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_UpperCamelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def _lowerCAmelCase ( _UpperCamelCase : Optional[int] ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE ={} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: _SCREAMING_SNAKE_CASE =bool(qa['answers']['text'] ) return qid_to_has_ans def _lowerCAmelCase ( _UpperCamelCase : Optional[int] ) -> Any: """simple docstring""" def remove_articles(_UpperCamelCase : List[Any] ): return ARTICLES_REGEX.sub(' ' , _UpperCamelCase ) def white_space_fix(_UpperCamelCase : List[Any] ): return " ".join(text.split() ) def remove_punc(_UpperCamelCase : Tuple ): _SCREAMING_SNAKE_CASE =set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_UpperCamelCase : List[Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_UpperCamelCase ) ) ) ) def _lowerCAmelCase ( _UpperCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" if not s: return [] return normalize_answer(_UpperCamelCase ).split() def _lowerCAmelCase ( _UpperCamelCase : Tuple , _UpperCamelCase : Tuple ) -> Optional[int]: """simple docstring""" return int(normalize_answer(_UpperCamelCase ) == normalize_answer(_UpperCamelCase ) ) def _lowerCAmelCase ( _UpperCamelCase : List[str] , _UpperCamelCase : Any ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =get_tokens(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =get_tokens(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =collections.Counter(_UpperCamelCase ) & collections.Counter(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =sum(common.values() ) if len(_UpperCamelCase ) == 0 or len(_UpperCamelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 _SCREAMING_SNAKE_CASE =1.0 * num_same / len(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =1.0 * num_same / len(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =(2 * precision * recall) / (precision + recall) return fa def _lowerCAmelCase ( _UpperCamelCase : int , _UpperCamelCase : Dict ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE ={} _SCREAMING_SNAKE_CASE ={} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: _SCREAMING_SNAKE_CASE =qa['id'] _SCREAMING_SNAKE_CASE =[t for t in qa['answers']['text'] if normalize_answer(_UpperCamelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string _SCREAMING_SNAKE_CASE =[''] if qid not in preds: print(f"Missing prediction for {qid}" ) continue _SCREAMING_SNAKE_CASE =preds[qid] # Take max over all gold answers _SCREAMING_SNAKE_CASE =max(compute_exact(_UpperCamelCase , _UpperCamelCase ) for a in gold_answers ) _SCREAMING_SNAKE_CASE =max(compute_fa(_UpperCamelCase , _UpperCamelCase ) for a in gold_answers ) return exact_scores, fa_scores def _lowerCAmelCase ( _UpperCamelCase : Tuple , _UpperCamelCase : List[str] , _UpperCamelCase : Any , _UpperCamelCase : Tuple ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE ={} for qid, s in scores.items(): _SCREAMING_SNAKE_CASE =na_probs[qid] > na_prob_thresh if pred_na: _SCREAMING_SNAKE_CASE =float(not qid_to_has_ans[qid] ) else: _SCREAMING_SNAKE_CASE =s return new_scores def _lowerCAmelCase ( _UpperCamelCase : str , _UpperCamelCase : int , _UpperCamelCase : Optional[Any]=None ) -> int: """simple docstring""" if not qid_list: _SCREAMING_SNAKE_CASE =len(_UpperCamelCase ) return collections.OrderedDict( [ ('exact', 1_00.0 * sum(exact_scores.values() ) / total), ('f1', 1_00.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: _SCREAMING_SNAKE_CASE =len(_UpperCamelCase ) return collections.OrderedDict( [ ('exact', 1_00.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_00.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def _lowerCAmelCase ( _UpperCamelCase : Optional[int] , _UpperCamelCase : List[Any] , _UpperCamelCase : Dict ) -> List[str]: """simple docstring""" for k in new_eval: _SCREAMING_SNAKE_CASE =new_eval[k] def _lowerCAmelCase ( _UpperCamelCase : int , _UpperCamelCase : List[str] , _UpperCamelCase : Any , _UpperCamelCase : int ) -> Any: """simple docstring""" plt.step(_UpperCamelCase , _UpperCamelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_UpperCamelCase , _UpperCamelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(_UpperCamelCase ) plt.savefig(_UpperCamelCase ) plt.clf() def _lowerCAmelCase ( _UpperCamelCase : Any , _UpperCamelCase : Optional[int] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : str , _UpperCamelCase : Tuple=None , _UpperCamelCase : Tuple=None ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE =sorted(_UpperCamelCase , key=lambda _UpperCamelCase : na_probs[k] ) _SCREAMING_SNAKE_CASE =0.0 _SCREAMING_SNAKE_CASE =1.0 _SCREAMING_SNAKE_CASE =0.0 _SCREAMING_SNAKE_CASE =[1.0] _SCREAMING_SNAKE_CASE =[0.0] _SCREAMING_SNAKE_CASE =0.0 for i, qid in enumerate(_UpperCamelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] _SCREAMING_SNAKE_CASE =true_pos / float(i + 1 ) _SCREAMING_SNAKE_CASE =true_pos / float(_UpperCamelCase ) if i == len(_UpperCamelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_UpperCamelCase ) recalls.append(_UpperCamelCase ) if out_image: plot_pr_curve(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) return {"ap": 1_00.0 * avg_prec} def _lowerCAmelCase ( _UpperCamelCase : Any , _UpperCamelCase : int , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : str , _UpperCamelCase : Tuple , _UpperCamelCase : Any ) -> Dict: """simple docstring""" if out_image_dir and not os.path.exists(_UpperCamelCase ): os.makedirs(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return _SCREAMING_SNAKE_CASE =make_precision_recall_eval( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , out_image=os.path.join(_UpperCamelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) _SCREAMING_SNAKE_CASE =make_precision_recall_eval( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , out_image=os.path.join(_UpperCamelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) _SCREAMING_SNAKE_CASE ={k: float(_UpperCamelCase ) for k, v in qid_to_has_ans.items()} _SCREAMING_SNAKE_CASE =make_precision_recall_eval( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , out_image=os.path.join(_UpperCamelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_UpperCamelCase , _UpperCamelCase , 'pr_exact' ) merge_eval(_UpperCamelCase , _UpperCamelCase , 'pr_f1' ) merge_eval(_UpperCamelCase , _UpperCamelCase , 'pr_oracle' ) def _lowerCAmelCase ( _UpperCamelCase : str , _UpperCamelCase : Dict , _UpperCamelCase : Tuple , _UpperCamelCase : str ) -> Union[str, Any]: """simple docstring""" if not qid_list: return _SCREAMING_SNAKE_CASE =[na_probs[k] for k in qid_list] _SCREAMING_SNAKE_CASE =np.ones_like(_UpperCamelCase ) / float(len(_UpperCamelCase ) ) plt.hist(_UpperCamelCase , weights=_UpperCamelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"Histogram of no-answer probability: {name}" ) plt.savefig(os.path.join(_UpperCamelCase , f"na_prob_hist_{name}.png" ) ) plt.clf() def _lowerCAmelCase ( _UpperCamelCase : Any , _UpperCamelCase : Dict , _UpperCamelCase : Any , _UpperCamelCase : List[Any] ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE =sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) _SCREAMING_SNAKE_CASE =num_no_ans _SCREAMING_SNAKE_CASE =cur_score _SCREAMING_SNAKE_CASE =0.0 _SCREAMING_SNAKE_CASE =sorted(_UpperCamelCase , key=lambda _UpperCamelCase : na_probs[k] ) for i, qid in enumerate(_UpperCamelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: _SCREAMING_SNAKE_CASE =scores[qid] else: if preds[qid]: _SCREAMING_SNAKE_CASE =-1 else: _SCREAMING_SNAKE_CASE =0 cur_score += diff if cur_score > best_score: _SCREAMING_SNAKE_CASE =cur_score _SCREAMING_SNAKE_CASE =na_probs[qid] return 1_00.0 * best_score / len(_UpperCamelCase ), best_thresh def _lowerCAmelCase ( _UpperCamelCase : str , _UpperCamelCase : str , _UpperCamelCase : Any , _UpperCamelCase : Optional[int] , _UpperCamelCase : Tuple , _UpperCamelCase : int ) -> List[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =find_best_thresh(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =find_best_thresh(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) _SCREAMING_SNAKE_CASE =best_exact _SCREAMING_SNAKE_CASE =exact_thresh _SCREAMING_SNAKE_CASE =best_fa _SCREAMING_SNAKE_CASE =fa_thresh def _lowerCAmelCase ( ) -> int: """simple docstring""" with open(OPTS.data_file ) as f: _SCREAMING_SNAKE_CASE =json.load(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =dataset_json['data'] with open(OPTS.pred_file ) as f: _SCREAMING_SNAKE_CASE =json.load(_UpperCamelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: _SCREAMING_SNAKE_CASE =json.load(_UpperCamelCase ) else: _SCREAMING_SNAKE_CASE ={k: 0.0 for k in preds} _SCREAMING_SNAKE_CASE =make_qid_to_has_ans(_UpperCamelCase ) # maps qid to True/False _SCREAMING_SNAKE_CASE =[k for k, v in qid_to_has_ans.items() if v] _SCREAMING_SNAKE_CASE =[k for k, v in qid_to_has_ans.items() if not v] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =get_raw_scores(_UpperCamelCase , _UpperCamelCase ) _SCREAMING_SNAKE_CASE =apply_no_ans_threshold(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , OPTS.na_prob_thresh ) _SCREAMING_SNAKE_CASE =apply_no_ans_threshold(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , OPTS.na_prob_thresh ) _SCREAMING_SNAKE_CASE =make_eval_dict(_UpperCamelCase , _UpperCamelCase ) if has_ans_qids: _SCREAMING_SNAKE_CASE =make_eval_dict(_UpperCamelCase , _UpperCamelCase , qid_list=_UpperCamelCase ) merge_eval(_UpperCamelCase , _UpperCamelCase , 'HasAns' ) if no_ans_qids: _SCREAMING_SNAKE_CASE =make_eval_dict(_UpperCamelCase , _UpperCamelCase , qid_list=_UpperCamelCase ) merge_eval(_UpperCamelCase , _UpperCamelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , OPTS.out_image_dir ) histogram_na_prob(_UpperCamelCase , _UpperCamelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_UpperCamelCase , _UpperCamelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_UpperCamelCase , _UpperCamelCase ) else: print(json.dumps(_UpperCamelCase , indent=2 ) ) if __name__ == "__main__": lowerCamelCase : List[Any] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
47
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCamelCase : Optional[int] = False class A__ ( unittest.TestCase ): pass @slow @require_torch_gpu class A__ ( unittest.TestCase ): def A ( self : Tuple ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) _SCREAMING_SNAKE_CASE =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) _SCREAMING_SNAKE_CASE =torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE =pipe( image=_a , generator=_a , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' , ).images _SCREAMING_SNAKE_CASE =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE =np.array([0.04_41, 0.04_69, 0.05_07, 0.05_75, 0.06_32, 0.06_50, 0.08_65, 0.09_09, 0.09_45] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
47
1
import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers SCREAMING_SNAKE_CASE :Optional[int] = 'python tqdm regex requests packaging filelock numpy tokenizers'.split() if sys.version_info < (3, 7): pkgs_to_check_at_runtime.append('dataclasses') if sys.version_info < (3, 8): pkgs_to_check_at_runtime.append('importlib_metadata') for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def UpperCAmelCase ( a_ , a_=None ) -> Any: """simple docstring""" require_version(deps[pkg] , a_ )
361
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = BertJapaneseTokenizer snake_case_ = False snake_case_ = True def UpperCamelCase_ ( self : List[Any] ): super().setUp() __A = [ "[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは", "世界", "##世界", "、", "##、", "。", "##。", ] __A = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def UpperCamelCase_ ( self : Union[str, Any] ,A : Optional[Any] ): __A = "こんにちは、世界。 \nこんばんは、世界。" __A = "こんにちは 、 世界 。 こんばんは 、 世界 。" return input_text, output_text def UpperCamelCase_ ( self : Any ,A : Optional[int] ): __A , __A = self.get_input_output_texts(A ) __A = tokenizer.encode(A ,add_special_tokens=A ) __A = tokenizer.decode(A ,clean_up_tokenization_spaces=A ) return text, ids def UpperCamelCase_ ( self : int ): pass # TODO add if relevant def UpperCamelCase_ ( self : int ): pass # TODO add if relevant def UpperCamelCase_ ( self : Optional[int] ): pass # TODO add if relevant def UpperCamelCase_ ( self : List[Any] ): __A = self.tokenizer_class(self.vocab_file ) __A = tokenizer.tokenize("こんにちは、世界。\nこんばんは、世界。" ) self.assertListEqual(A ,["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) def UpperCamelCase_ ( self : int ): __A = self.tokenizer_class(self.vocab_file ,word_tokenizer_type="mecab" ) self.assertIsNotNone(A ) __A = "こんにちは、世界。\nこんばんは、世界。" __A = tokenizer.tokenize(A ) self.assertListEqual(A ,["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) __A = os.path.join(self.tmpdirname ,"tokenizer.bin" ) with open(A ,"wb" ) as handle: pickle.dump(A ,A ) with open(A ,"rb" ) as handle: __A = pickle.load(A ) __A = tokenizer_new.tokenize(A ) self.assertListEqual(A ,A ) def UpperCamelCase_ ( self : Any ): __A = MecabTokenizer(mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] ,) def UpperCamelCase_ ( self : List[str] ): try: __A = MecabTokenizer(mecab_dic="unidic_lite" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] ,) def UpperCamelCase_ ( self : Tuple ): try: __A = MecabTokenizer(mecab_dic="unidic" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] ,) def UpperCamelCase_ ( self : Tuple ): __A = MecabTokenizer(do_lower_case=A ,mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップルストア", "で", "iphone", "8", "が", "発売", "さ", "れ", "た", "。"] ,) def UpperCamelCase_ ( self : Any ): try: __A = MecabTokenizer( do_lower_case=A ,normalize_text=A ,mecab_option="-d /usr/local/lib/mecab/dic/jumandic" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "\u3000", "。"] ,) def UpperCamelCase_ ( self : int ): __A = MecabTokenizer(normalize_text=A ,mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", " ", "。"] ,) @require_sudachi def UpperCamelCase_ ( self : int ): __A = self.tokenizer_class(self.vocab_file ,word_tokenizer_type="sudachi" ) self.assertIsNotNone(A ) __A = "こんにちは、世界。\nこんばんは、世界。" __A = tokenizer.tokenize(A ) self.assertListEqual(A ,["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) __A = os.path.join(self.tmpdirname ,"tokenizer.bin" ) with open(A ,"wb" ) as handle: pickle.dump(A ,A ) with open(A ,"rb" ) as handle: __A = pickle.load(A ) __A = tokenizer_new.tokenize(A ) self.assertListEqual(A ,A ) @require_sudachi def UpperCamelCase_ ( self : Optional[int] ): __A = SudachiTokenizer(sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,[" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] ,) @require_sudachi def UpperCamelCase_ ( self : List[Any] ): __A = SudachiTokenizer(sudachi_dict_type="core" ,sudachi_split_mode="A" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) ,["外国", "人", "参政", "権"] ) @require_sudachi def UpperCamelCase_ ( self : int ): __A = SudachiTokenizer(sudachi_dict_type="core" ,sudachi_split_mode="B" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) ,["外国人", "参政権"] ) @require_sudachi def UpperCamelCase_ ( self : Tuple ): __A = SudachiTokenizer(sudachi_dict_type="core" ,sudachi_split_mode="C" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) ,["外国人参政権"] ) @require_sudachi def UpperCamelCase_ ( self : int ): __A = SudachiTokenizer(do_lower_case=A ,sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,[" ", "\t", "アップル", "ストア", "で", "iphone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] ,) @require_sudachi def UpperCamelCase_ ( self : List[str] ): __A = SudachiTokenizer(normalize_text=A ,sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,[" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", "\u3000", "。", " ", " "] ,) @require_sudachi def UpperCamelCase_ ( self : str ): __A = SudachiTokenizer(trim_whitespace=A ,sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] ,) @require_jumanpp def UpperCamelCase_ ( self : Optional[Any] ): __A = self.tokenizer_class(self.vocab_file ,word_tokenizer_type="jumanpp" ) self.assertIsNotNone(A ) __A = "こんにちは、世界。\nこんばんは、世界。" __A = tokenizer.tokenize(A ) self.assertListEqual(A ,["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) __A = os.path.join(self.tmpdirname ,"tokenizer.bin" ) with open(A ,"wb" ) as handle: pickle.dump(A ,A ) with open(A ,"rb" ) as handle: __A = pickle.load(A ) __A = tokenizer_new.tokenize(A ) self.assertListEqual(A ,A ) @require_jumanpp def UpperCamelCase_ ( self : int ): __A = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] ,) @require_jumanpp def UpperCamelCase_ ( self : str ): __A = JumanppTokenizer(do_lower_case=A ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップル", "ストア", "で", "iphone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] ,) @require_jumanpp def UpperCamelCase_ ( self : Any ): __A = JumanppTokenizer(normalize_text=A ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["ア", "ッ", "フ", "゚", "ル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] ,) @require_jumanpp def UpperCamelCase_ ( self : List[str] ): __A = JumanppTokenizer(trim_whitespace=A ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) ,["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "。"] ,) @require_jumanpp def UpperCamelCase_ ( self : Dict ): __A = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("ありがとうございますm(_ _)m見つけるのが大変です。" ) ,["ありがとう", "ございます", "m(_ _)m", "見つける", "の", "が", "大変です", "。"] ,) def UpperCamelCase_ ( self : str ): __A = ["[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは"] __A = {} for i, token in enumerate(A ): __A = i __A = WordpieceTokenizer(vocab=A ,unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) ,[] ) self.assertListEqual(tokenizer.tokenize("こんにちは" ) ,["こんにちは"] ) self.assertListEqual(tokenizer.tokenize("こんばんは" ) ,["こん", "##ばんは"] ) self.assertListEqual(tokenizer.tokenize("こんばんは こんばんにちは こんにちは" ) ,["こん", "##ばんは", "[UNK]", "こんにちは"] ) def UpperCamelCase_ ( self : Any ): __A = BertJapaneseTokenizer.from_pretrained("nlp-waseda/roberta-base-japanese-with-auto-jumanpp" ) __A = tokenizer.subword_tokenizer __A = subword_tokenizer.tokenize("国境 の 長い トンネル を 抜ける と 雪国 であった 。" ) self.assertListEqual(A ,["▁国境", "▁の", "▁長い", "▁トンネル", "▁を", "▁抜ける", "▁と", "▁雪", "国", "▁であった", "▁。"] ) __A = subword_tokenizer.tokenize("こんばんは こんばん にち は こんにちは" ) self.assertListEqual(A ,["▁こん", "ばん", "は", "▁こん", "ばん", "▁に", "ち", "▁は", "▁こんにちは"] ) def UpperCamelCase_ ( self : int ): __A = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese" ) __A = tokenizer.encode("ありがとう。" ,add_special_tokens=A ) __A = tokenizer.encode("どういたしまして。" ,add_special_tokens=A ) __A = tokenizer.build_inputs_with_special_tokens(A ) __A = tokenizer.build_inputs_with_special_tokens(A ,A ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = BertJapaneseTokenizer snake_case_ = False def UpperCamelCase_ ( self : Any ): super().setUp() __A = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] __A = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def UpperCamelCase_ ( self : int ,**A : str ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname ,subword_tokenizer_type="character" ,**A ) def UpperCamelCase_ ( self : List[str] ,A : str ): __A = "こんにちは、世界。 \nこんばんは、世界。" __A = "こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。" return input_text, output_text def UpperCamelCase_ ( self : str ): pass # TODO add if relevant def UpperCamelCase_ ( self : Optional[Any] ): pass # TODO add if relevant def UpperCamelCase_ ( self : Any ): pass # TODO add if relevant def UpperCamelCase_ ( self : str ): __A = self.tokenizer_class(self.vocab_file ,subword_tokenizer_type="character" ) __A = tokenizer.tokenize("こんにちは、世界。 \nこんばんは、世界。" ) self.assertListEqual( A ,["こ", "ん", "に", "ち", "は", "、", "世", "界", "。", "こ", "ん", "ば", "ん", "は", "、", "世", "界", "。"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A ) ,[3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def UpperCamelCase_ ( self : Union[str, Any] ): __A = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] __A = {} for i, token in enumerate(A ): __A = i __A = CharacterTokenizer(vocab=A ,unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) ,[] ) self.assertListEqual(tokenizer.tokenize("こんにちは" ) ,["こ", "ん", "に", "ち", "は"] ) self.assertListEqual(tokenizer.tokenize("こんにちほ" ) ,["こ", "ん", "に", "ち", "[UNK]"] ) def UpperCamelCase_ ( self : Tuple ): __A = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese-char" ) __A = tokenizer.encode("ありがとう。" ,add_special_tokens=A ) __A = tokenizer.encode("どういたしまして。" ,add_special_tokens=A ) __A = tokenizer.build_inputs_with_special_tokens(A ) __A = tokenizer.build_inputs_with_special_tokens(A ,A ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : int ): __A = "cl-tohoku/bert-base-japanese" __A = AutoTokenizer.from_pretrained(A ) self.assertIsInstance(A ,A ) class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Tuple ): __A = "cl-tohoku/bert-base-japanese" with self.assertLogs("transformers" ,level="WARNING" ) as cm: BertTokenizer.from_pretrained(A ) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from." ) ) __A = "bert-base-cased" with self.assertLogs("transformers" ,level="WARNING" ) as cm: BertJapaneseTokenizer.from_pretrained(A ) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from." ) )
124
0
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' with open(lowerCAmelCase_ ) as metadata_file: __SCREAMING_SNAKE_CASE = json.load(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = LukeConfig(use_entity_aware_attention=lowerCAmelCase_ , **metadata["model_config"] ) # Load in the weights from the checkpoint_path __SCREAMING_SNAKE_CASE = torch.load(lowerCAmelCase_ , map_location="cpu" ) # Load the entity vocab file __SCREAMING_SNAKE_CASE = load_entity_vocab(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = RobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks __SCREAMING_SNAKE_CASE = AddedToken("<ent>" , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = AddedToken("<ent2>" , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(f"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowerCAmelCase_ ) with open(os.path.join(lowerCAmelCase_ , LukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(lowerCAmelCase_ , lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = LukeTokenizer.from_pretrained(lowerCAmelCase_ ) # Initialize the embeddings of the special tokens __SCREAMING_SNAKE_CASE = state_dict["embeddings.word_embeddings.weight"] __SCREAMING_SNAKE_CASE = word_emb[tokenizer.convert_tokens_to_ids(["@"] )[0]].unsqueeze(0 ) __SCREAMING_SNAKE_CASE = word_emb[tokenizer.convert_tokens_to_ids(["#"] )[0]].unsqueeze(0 ) __SCREAMING_SNAKE_CASE = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __SCREAMING_SNAKE_CASE = f"""encoder.layer.{layer_index}.attention.self.""" __SCREAMING_SNAKE_CASE = state_dict[prefix + matrix_name] __SCREAMING_SNAKE_CASE = state_dict[prefix + matrix_name] __SCREAMING_SNAKE_CASE = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __SCREAMING_SNAKE_CASE = state_dict["entity_embeddings.entity_embeddings.weight"] __SCREAMING_SNAKE_CASE = entity_emb[entity_vocab["[MASK]"]] __SCREAMING_SNAKE_CASE = LukeModel(config=lowerCAmelCase_ ).eval() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) if not (len(lowerCAmelCase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(f"""Missing keys {', '.join(lowerCAmelCase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith("entity_predictions" ) or key.startswith("lm_head" ) for key in unexpected_keys )): raise ValueError( "Unexpected keys" f""" {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}""" ) # Check outputs __SCREAMING_SNAKE_CASE = LukeTokenizer.from_pretrained(lowerCAmelCase_ , task="entity_classification" ) __SCREAMING_SNAKE_CASE = ( "Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the" " new world number one avoid a humiliating second- round exit at Wimbledon ." ) __SCREAMING_SNAKE_CASE = (39, 42) __SCREAMING_SNAKE_CASE = tokenizer(lowerCAmelCase_ , entity_spans=[span] , add_prefix_space=lowerCAmelCase_ , return_tensors="pt" ) __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase_ ) # Verify word hidden states if model_size == "large": __SCREAMING_SNAKE_CASE = torch.Size((1, 42, 1024) ) __SCREAMING_SNAKE_CASE = torch.tensor( [[0.0133, 0.0865, 0.0095], [0.3093, -0.2576, -0.7418], [-0.1720, -0.2117, -0.2869]] ) else: # base __SCREAMING_SNAKE_CASE = torch.Size((1, 42, 768) ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.0037, 0.1368, -0.0091], [0.1099, 0.3329, -0.1095], [0.0765, 0.5335, 0.1179]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCAmelCase_ , atol=1E-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __SCREAMING_SNAKE_CASE = torch.Size((1, 1, 1024) ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.0466, -0.0106, -0.0179]] ) else: # base __SCREAMING_SNAKE_CASE = torch.Size((1, 1, 768) ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.1457, 0.1044, 0.0174]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( f"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" f""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowerCAmelCase_ , atol=1E-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(lowerCAmelCase_ ) ) model.save_pretrained(lowerCAmelCase_ ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = {} with open(lowerCAmelCase_ , "r" , encoding="utf-8" ) as f: for index, line in enumerate(lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = line.rstrip().split("\t" ) __SCREAMING_SNAKE_CASE = index return entity_vocab if __name__ == "__main__": a__ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) a__ : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
54
'''simple docstring''' from PIL import Image def _a( UpperCamelCase__ : Image, UpperCamelCase__ : float ): '''simple docstring''' def brightness(UpperCamelCase__ : int ) -> float: return 1_2_8 + level + (c - 1_2_8) if not -2_5_5.0 <= level <= 2_5_5.0: raise ValueError('''level must be between -255.0 (black) and 255.0 (white)''' ) return img.point(UpperCamelCase__ ) if __name__ == "__main__": # Load image with Image.open('image_data/lena.jpg') as img: # Change brightness to 100 a_ = change_brightness(img, 1_0_0) brigt_img.save('image_data/lena_brightness.png', format='png')
152
0
"""simple docstring""" import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __lowerCamelCase ( A__ ): '''simple docstring''' def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : List[str] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(a_ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(a_ , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(a_ , "num_attention_heads" ) ) class __lowerCamelCase : '''simple docstring''' def __init__( self : Optional[Any] , a_ : int , a_ : Optional[int]=13 , a_ : int=32 , a_ : Optional[int]=2 , a_ : Union[str, Any]=3 , a_ : Optional[Any]=6_40 , a_ : Any=4 , a_ : Union[str, Any]="silu" , a_ : Optional[int]=3 , a_ : Tuple=32 , a_ : Optional[Any]=0.1 , a_ : Optional[int]=0.1 , a_ : int=0.1 , a_ : Tuple=0.02 , a_ : int=True , a_ : Dict=True , a_ : Any=10 , a_ : List[str]=None , ): lowerCAmelCase_ : Optional[int] = parent lowerCAmelCase_ : Dict = batch_size lowerCAmelCase_ : Any = image_size lowerCAmelCase_ : List[str] = patch_size lowerCAmelCase_ : int = num_channels lowerCAmelCase_ : List[Any] = last_hidden_size lowerCAmelCase_ : Dict = num_attention_heads lowerCAmelCase_ : Optional[int] = hidden_act lowerCAmelCase_ : Optional[Any] = conv_kernel_size lowerCAmelCase_ : Union[str, Any] = output_stride lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase_ : List[Any] = classifier_dropout_prob lowerCAmelCase_ : List[str] = use_labels lowerCAmelCase_ : str = is_training lowerCAmelCase_ : Dict = num_labels lowerCAmelCase_ : Dict = initializer_range lowerCAmelCase_ : Any = scope def lowerCamelCase ( self : str ): lowerCAmelCase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase_ : int = None lowerCAmelCase_ : Optional[Any] = None if self.use_labels: lowerCAmelCase_ : int = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase_ : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCAmelCase_ : int = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase ( self : str ): return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase ( self : Union[str, Any] , a_ : Optional[Any] , a_ : int , a_ : str , a_ : str ): lowerCAmelCase_ : int = MobileViTModel(config=a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : List[Any] = model(a_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase ( self : str , a_ : Tuple , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : int ): lowerCAmelCase_ : str = self.num_labels lowerCAmelCase_ : Tuple = MobileViTForImageClassification(a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : Optional[int] = model(a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self : str , a_ : Any , a_ : Optional[Any] , a_ : str , a_ : Optional[int] ): lowerCAmelCase_ : Optional[Any] = self.num_labels lowerCAmelCase_ : List[Any] = MobileViTForSemanticSegmentation(a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : List[str] = model(a_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowerCAmelCase_ : List[Any] = model(a_ , labels=a_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase ( self : str ): lowerCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() lowerCAmelCase_ : Union[str, Any] = config_and_inputs lowerCAmelCase_ : Any = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( A__ , A__ , unittest.TestCase ): '''simple docstring''' a_ : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) a_ : Tuple = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) a_ : Optional[int] = False a_ : int = False a_ : Any = False a_ : List[Any] = False def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : Optional[int] = MobileViTModelTester(self ) lowerCAmelCase_ : int = MobileViTConfigTester(self , config_class=a_ , has_text_modality=a_ ) def lowerCamelCase ( self : Optional[int] ): self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def lowerCamelCase ( self : Union[str, Any] ): pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def lowerCamelCase ( self : Any ): pass @unittest.skip(reason="MobileViT does not output attentions" ) def lowerCamelCase ( self : Tuple ): pass def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ : Union[str, Any] = model_class(a_ ) lowerCAmelCase_ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase_ : Tuple = [*signature.parameters.keys()] lowerCAmelCase_ : List[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , a_ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def lowerCamelCase ( self : Any ): pass def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def lowerCamelCase ( self : str ): def check_hidden_states_output(a_ : int , a_ : Union[str, Any] , a_ : Optional[int] ): lowerCAmelCase_ : Union[str, Any] = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase_ : List[Any] = model(**self._prepare_for_class(a_ , a_ ) ) lowerCAmelCase_ : Any = outputs.hidden_states lowerCAmelCase_ : int = 5 self.assertEqual(len(a_ ) , a_ ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. lowerCAmelCase_ : int = 2 for i in range(len(a_ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ : List[Any] = True check_hidden_states_output(a_ , a_ , a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase_ : str = True check_hidden_states_output(a_ , a_ , a_ ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*a_ ) @slow def lowerCamelCase ( self : List[str] ): for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = MobileViTModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def __lowerCamelCase ( ) -> Dict: """simple docstring""" lowerCAmelCase_ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCamelCase ( self : List[Any] ): return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : int = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(a_ ) lowerCAmelCase_ : Dict = self.default_image_processor lowerCAmelCase_ : List[str] = prepare_img() lowerCAmelCase_ : Union[str, Any] = image_processor(images=a_ , return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): lowerCAmelCase_ : str = model(**a_ ) # verify the logits lowerCAmelCase_ : Union[str, Any] = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , a_ ) lowerCAmelCase_ : Union[str, Any] = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a_ , atol=1e-4 ) ) @slow def lowerCamelCase ( self : Union[str, Any] ): lowerCAmelCase_ : Tuple = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCAmelCase_ : Optional[int] = model.to(a_ ) lowerCAmelCase_ : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCAmelCase_ : int = prepare_img() lowerCAmelCase_ : Tuple = image_processor(images=a_ , return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): lowerCAmelCase_ : Union[str, Any] = model(**a_ ) lowerCAmelCase_ : Any = outputs.logits # verify the logits lowerCAmelCase_ : List[Any] = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , a_ ) lowerCAmelCase_ : Dict = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=a_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , a_ , atol=1e-4 ) ) @slow def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : Union[str, Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCAmelCase_ : Tuple = model.to(a_ ) lowerCAmelCase_ : Tuple = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) lowerCAmelCase_ : Tuple = prepare_img() lowerCAmelCase_ : Optional[int] = image_processor(images=a_ , return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): lowerCAmelCase_ : List[str] = model(**a_ ) lowerCAmelCase_ : Optional[Any] = outputs.logits.detach().cpu() lowerCAmelCase_ : Any = image_processor.post_process_semantic_segmentation(outputs=a_ , target_sizes=[(50, 60)] ) lowerCAmelCase_ : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , a_ ) lowerCAmelCase_ : Union[str, Any] = image_processor.post_process_semantic_segmentation(outputs=a_ ) lowerCAmelCase_ : List[Any] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , a_ )
363
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase__ = random.Random() if is_torch_available(): import torch def __lowerCamelCase ( __UpperCamelCase , __UpperCamelCase=1.0 , __UpperCamelCase=None , __UpperCamelCase=None ) -> Dict: """simple docstring""" if rng is None: lowerCAmelCase_ : int = global_rng lowerCAmelCase_ : Dict = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Dict , a_ : Dict , a_ : Dict=7 , a_ : int=4_00 , a_ : Union[str, Any]=20_00 , a_ : Any=1 , a_ : Optional[int]=0.0 , a_ : str=1_60_00 , a_ : Optional[int]=True , a_ : Dict=True , ): lowerCAmelCase_ : Tuple = parent lowerCAmelCase_ : Union[str, Any] = batch_size lowerCAmelCase_ : Optional[int] = min_seq_length lowerCAmelCase_ : List[Any] = max_seq_length lowerCAmelCase_ : Optional[Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCAmelCase_ : Dict = feature_size lowerCAmelCase_ : Tuple = padding_value lowerCAmelCase_ : int = sampling_rate lowerCAmelCase_ : str = return_attention_mask lowerCAmelCase_ : Union[str, Any] = do_normalize def lowerCamelCase ( self : Dict ): return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def lowerCamelCase ( self : List[Any] , a_ : List[Any]=False , a_ : Optional[int]=False ): def _flatten(a_ : Optional[Any] ): return list(itertools.chain(*a_ ) ) if equal_length: lowerCAmelCase_ : Optional[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size lowerCAmelCase_ : Any = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase_ : List[Any] = [np.asarray(a_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowerCamelCase ( A__ , unittest.TestCase ): '''simple docstring''' a_ : Tuple = ASTFeatureExtractor def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Optional[int] = ASTFeatureExtractionTester(self ) def lowerCamelCase ( self : Tuple ): # Tests that all call wrap to encode_plus and batch_encode_plus lowerCAmelCase_ : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase_ : Tuple = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowerCAmelCase_ : str = [np.asarray(a_ ) for speech_input in speech_inputs] # Test not batched input lowerCAmelCase_ : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values lowerCAmelCase_ : Union[str, Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(a_ , a_ , atol=1e-3 ) ) # Test batched lowerCAmelCase_ : Tuple = feat_extract(a_ , padding=a_ , return_tensors="np" ).input_values lowerCAmelCase_ : int = feat_extract(a_ , padding=a_ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(a_ , a_ ): self.assertTrue(np.allclose(a_ , a_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. lowerCAmelCase_ : Tuple = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] lowerCAmelCase_ : Union[str, Any] = np.asarray(a_ ) lowerCAmelCase_ : str = feat_extract(a_ , return_tensors="np" ).input_values lowerCAmelCase_ : List[Any] = feat_extract(a_ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(a_ , a_ ): self.assertTrue(np.allclose(a_ , a_ , atol=1e-3 ) ) @require_torch def lowerCamelCase ( self : List[str] ): import torch lowerCAmelCase_ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase_ : Tuple = np.random.rand(1_00 ).astype(np.floataa ) lowerCAmelCase_ : List[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCAmelCase_ : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) lowerCAmelCase_ : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def lowerCamelCase ( self : List[Any] , a_ : List[str] ): from datasets import load_dataset lowerCAmelCase_ : Union[str, Any] = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech lowerCAmelCase_ : Optional[int] = ds.sort("id" ).select(range(a_ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] @require_torch def lowerCamelCase ( self : str ): # fmt: off lowerCAmelCase_ : Tuple = torch.tensor( [-0.9894, -1.2776, -0.9066, -1.2776, -0.9349, -1.2609, -1.0386, -1.2776, -1.1561, -1.2776, -1.2052, -1.2723, -1.2190, -1.2132, -1.2776, -1.1133, -1.1953, -1.1343, -1.1584, -1.2203, -1.1770, -1.2474, -1.2381, -1.1936, -0.9270, -0.8317, -0.8049, -0.7706, -0.7565, -0.7869] ) # fmt: on lowerCAmelCase_ : Dict = self._load_datasamples(1 ) lowerCAmelCase_ : Union[str, Any] = ASTFeatureExtractor() lowerCAmelCase_ : int = feature_extractor(a_ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 10_24, 1_28) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , a_ , atol=1e-4 ) )
161
0
from math import sqrt def lowerCamelCase_ ( _a : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" UpperCAmelCase_ : str = True # 0 and 1 are none primes. if number <= 1: UpperCAmelCase_ : Any = False for divisor in range(2 , int(round(sqrt(_lowerCAmelCase ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: UpperCAmelCase_ : Tuple = False break # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'status' must been from type bool" return status def lowerCamelCase_ ( _a : Tuple ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N UpperCAmelCase_ : Optional[int] = list(range(2 , n + 1 ) ) UpperCAmelCase_ : List[str] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(_lowerCAmelCase ) ): for j in range(i + 1 , len(_lowerCAmelCase ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): UpperCAmelCase_ : Optional[int] = 0 # filters actual prime numbers. UpperCAmelCase_ : Dict = [x for x in begin_list if x != 0] # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def lowerCamelCase_ ( _a : List[Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n > 2), "'N' must been an int and > 2" UpperCAmelCase_ : str = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(_lowerCAmelCase ): ans.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def lowerCamelCase_ ( _a : Optional[int] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and number >= 0, "'number' must been an int and >= 0" UpperCAmelCase_ : Optional[Any] = [] # this list will be returns of the function. # potential prime number factors. UpperCAmelCase_ : Dict = 2 UpperCAmelCase_ : Dict = number if number == 0 or number == 1: ans.append(_lowerCAmelCase ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(_lowerCAmelCase ): while quotient != 1: if is_prime(_lowerCAmelCase ) and (quotient % factor == 0): ans.append(_lowerCAmelCase ) quotient /= factor else: factor += 1 else: ans.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def lowerCamelCase_ ( _a : Any ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" UpperCAmelCase_ : Optional[Any] = 0 # prime factorization of 'number' UpperCAmelCase_ : Tuple = prime_factorization(_lowerCAmelCase ) UpperCAmelCase_ : Dict = max(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type int" return ans def lowerCamelCase_ ( _a : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" UpperCAmelCase_ : Tuple = 0 # prime factorization of 'number' UpperCAmelCase_ : List[str] = prime_factorization(_lowerCAmelCase ) UpperCAmelCase_ : List[Any] = min(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type int" return ans def lowerCamelCase_ ( _a : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'number' must been an int" assert isinstance(number % 2 == 0 , _lowerCAmelCase ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase_ ( _a : List[str] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'number' must been an int" assert isinstance(number % 2 != 0 , _lowerCAmelCase ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase_ ( _a : Union[str, Any] ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (number > 2) and is_even(_lowerCAmelCase ) ), "'number' must been an int, even and > 2" UpperCAmelCase_ : Tuple = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' UpperCAmelCase_ : int = get_prime_numbers(_lowerCAmelCase ) UpperCAmelCase_ : Tuple = len(_lowerCAmelCase ) # run variable for while-loops. UpperCAmelCase_ : Optional[int] = 0 UpperCAmelCase_ : List[Any] = None # exit variable. for break up the loops UpperCAmelCase_ : Union[str, Any] = True while i < len_pn and loop: UpperCAmelCase_ : Dict = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: UpperCAmelCase_ : List[Any] = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (len(_lowerCAmelCase ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase_ ( _a : str , _a : Tuple ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." UpperCAmelCase_ : Any = 0 while numbera != 0: UpperCAmelCase_ : Union[str, Any] = numbera % numbera UpperCAmelCase_ : List[Any] = numbera UpperCAmelCase_ : List[Any] = rest # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase_ ( _a : Tuple , _a : Optional[int] ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." UpperCAmelCase_ : Union[str, Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' UpperCAmelCase_ : Tuple = prime_factorization(_lowerCAmelCase ) UpperCAmelCase_ : int = prime_factorization(_lowerCAmelCase ) elif numbera == 1 or numbera == 1: UpperCAmelCase_ : Optional[Any] = [] UpperCAmelCase_ : Tuple = [] UpperCAmelCase_ : List[Any] = max(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase_ : str = 0 UpperCAmelCase_ : List[str] = 0 UpperCAmelCase_ : Any = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: UpperCAmelCase_ : Optional[int] = prime_fac_a.count(_lowerCAmelCase ) UpperCAmelCase_ : Dict = prime_fac_a.count(_lowerCAmelCase ) for _ in range(max(_lowerCAmelCase , _lowerCAmelCase ) ): ans *= n else: UpperCAmelCase_ : str = prime_fac_a.count(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ): ans *= n done.append(_lowerCAmelCase ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: UpperCAmelCase_ : int = prime_fac_a.count(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ): ans *= n done.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase_ ( _a : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'number' must been a positive int" UpperCAmelCase_ : Optional[int] = 0 UpperCAmelCase_ : Dict = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(_lowerCAmelCase ): ans += 1 # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and is_prime( _lowerCAmelCase ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase_ ( _a : List[Any] , _a : Union[str, Any] ): '''simple docstring''' assert ( is_prime(_lowerCAmelCase ) and is_prime(_lowerCAmelCase ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" UpperCAmelCase_ : Optional[Any] = p_number_a + 1 # jump to the next number UpperCAmelCase_ : List[Any] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(_lowerCAmelCase ): number += 1 while number < p_number_a: ans.append(_lowerCAmelCase ) number += 1 # fetch the next prime number. while not is_prime(_lowerCAmelCase ): number += 1 # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ans[0] != p_number_a and ans[len(_lowerCAmelCase ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase_ ( _a : Union[str, Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 1), "'n' must been int and >= 1" UpperCAmelCase_ : Optional[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(_lowerCAmelCase ) # precondition assert ans[0] == 1 and ans[len(_lowerCAmelCase ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase_ ( _a : Union[str, Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number > 1 ), "'number' must been an int and >= 1" UpperCAmelCase_ : List[Any] = get_divisors(_lowerCAmelCase ) # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (divisors[0] == 1) and (divisors[len(_lowerCAmelCase ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase_ ( _a : Dict , _a : Optional[Any] ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. UpperCAmelCase_ : Optional[int] = gcd(abs(_lowerCAmelCase ) , abs(_lowerCAmelCase ) ) # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase_ ( _a : List[str] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'n' must been a int and >= 0" UpperCAmelCase_ : Optional[Any] = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCamelCase_ ( _a : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'n' must been an int and >= 0" UpperCAmelCase_ : Tuple = 0 UpperCAmelCase_ : Optional[Any] = 1 UpperCAmelCase_ : Optional[int] = 1 # this will be return for _ in range(n - 1 ): UpperCAmelCase_ : Any = ans ans += fiba UpperCAmelCase_ : List[str] = tmp return ans
345
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def A_ ( _lowerCAmelCase : Dict ): """simple docstring""" if ( (cp >= 0x4e00 and cp <= 0x9fff) or (cp >= 0x3400 and cp <= 0x4dbf) # or (cp >= 0x2_0000 and cp <= 0x2_a6df) # or (cp >= 0x2_a700 and cp <= 0x2_b73f) # or (cp >= 0x2_b740 and cp <= 0x2_b81f) # or (cp >= 0x2_b820 and cp <= 0x2_ceaf) # or (cp >= 0xf900 and cp <= 0xfaff) or (cp >= 0x2_f800 and cp <= 0x2_fa1f) # ): # return True return False def A_ ( _lowerCAmelCase : str ): """simple docstring""" for char in word: _a = ord(_lowerCAmelCase ) if not _is_chinese_char(_lowerCAmelCase ): return 0 return 1 def A_ ( _lowerCAmelCase : List[str] ): """simple docstring""" _a = set() for token in tokens: _a = len(_lowerCAmelCase ) > 1 and is_chinese(_lowerCAmelCase ) if chinese_word: word_set.add(_lowerCAmelCase ) _a = list(_lowerCAmelCase ) return word_list def A_ ( _lowerCAmelCase : List[str], _lowerCAmelCase : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a = max([len(_lowerCAmelCase ) for w in chinese_word_set] ) _a = bert_tokens _a , _a = 0, len(_lowerCAmelCase ) while start < end: _a = True if is_chinese(bert_word[start] ): _a = min(end - start, _lowerCAmelCase ) for i in range(_lowerCAmelCase, 1, -1 ): _a = ''''''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1, start + i ): _a = '''##''' + bert_word[j] _a = start + i _a = False break if single_word: start += 1 return bert_word def A_ ( _lowerCAmelCase : List[str], _lowerCAmelCase : LTP, _lowerCAmelCase : BertTokenizer ): """simple docstring""" _a = [] for i in range(0, len(_lowerCAmelCase ), 1_00 ): _a = ltp_tokenizer.pipeline(lines[i : i + 1_00], tasks=['''cws'''] ).cws _a = [get_chinese_word(_lowerCAmelCase ) for r in res] ltp_res.extend(_lowerCAmelCase ) assert len(_lowerCAmelCase ) == len(_lowerCAmelCase ) _a = [] for i in range(0, len(_lowerCAmelCase ), 1_00 ): _a = bert_tokenizer(lines[i : i + 1_00], add_special_tokens=_lowerCAmelCase, truncation=_lowerCAmelCase, max_length=5_12 ) bert_res.extend(res['''input_ids'''] ) assert len(_lowerCAmelCase ) == len(_lowerCAmelCase ) _a = [] for input_ids, chinese_word in zip(_lowerCAmelCase, _lowerCAmelCase ): _a = [] for id in input_ids: _a = bert_tokenizer._convert_id_to_token(_lowerCAmelCase ) input_tokens.append(_lowerCAmelCase ) _a = add_sub_symbol(_lowerCAmelCase, _lowerCAmelCase ) _a = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(_lowerCAmelCase ): if token[:2] == "##": _a = token[2:] # save chinese tokens' pos if len(_lowerCAmelCase ) == 1 and _is_chinese_char(ord(_lowerCAmelCase ) ): ref_id.append(_lowerCAmelCase ) ref_ids.append(_lowerCAmelCase ) assert len(_lowerCAmelCase ) == len(_lowerCAmelCase ) return ref_ids def A_ ( _lowerCAmelCase : Any ): """simple docstring""" with open(args.file_name, '''r''', encoding='''utf-8''' ) as f: _a = f.readlines() _a = [line.strip() for line in data if len(_lowerCAmelCase ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a = LTP(args.ltp ) # faster in GPU device _a = BertTokenizer.from_pretrained(args.bert ) _a = prepare_ref(_lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase ) with open(args.save_path, '''w''', encoding='''utf-8''' ) as f: _a = [json.dumps(_lowerCAmelCase ) + '''\n''' for ref in ref_ids] f.writelines(_lowerCAmelCase ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser(description='''prepare_chinese_ref''') parser.add_argument( '''--file_name''', required=False, type=str, default='''./resources/chinese-demo.txt''', help='''file need process, same as training data in lm''', ) parser.add_argument( '''--ltp''', required=False, type=str, default='''./resources/ltp''', help='''resources for LTP tokenizer, usually a path''', ) parser.add_argument( '''--bert''', required=False, type=str, default='''./resources/robert''', help='''resources for Bert tokenizer''', ) parser.add_argument( '''--save_path''', required=False, type=str, default='''./resources/ref.txt''', help='''path to save res''', ) __snake_case = parser.parse_args() main(args)
320
0
"""simple docstring""" import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Dict = DownBlockaD # noqa F405 __magic_name__ :Tuple = """down""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = [-0.02_32, -0.98_69, 0.80_54, -0.06_37, -0.16_88, -1.42_64, 0.44_70, -1.33_94, 0.09_04] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Dict = ResnetDownsampleBlockaD # noqa F405 __magic_name__ :Dict = """down""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [0.07_10, 0.24_10, -0.73_20, -1.07_57, -1.13_43, 0.35_40, -0.01_33, -0.25_76, 0.09_48] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Tuple = AttnDownBlockaD # noqa F405 __magic_name__ :List[str] = """down""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = [0.06_36, 0.89_64, -0.62_34, -1.01_31, 0.08_44, 0.49_35, 0.34_37, 0.09_11, -0.29_57] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Union[str, Any] = CrossAttnDownBlockaD # noqa F405 __magic_name__ :Optional[int] = """down""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :int = super().prepare_init_args_and_inputs_for_common() lowerCAmelCase__ :Optional[int] = 3_2 return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = [0.22_38, -0.73_96, -0.22_55, -0.38_29, 0.19_25, 1.16_65, 0.06_03, -0.72_95, 0.19_83] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :str = SimpleCrossAttnDownBlockaD # noqa F405 __magic_name__ :Optional[int] = """down""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_encoder_hidden_states=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :Optional[Any] = super().prepare_init_args_and_inputs_for_common() lowerCAmelCase__ :List[str] = 3_2 return init_dict, inputs_dict @unittest.skipIf(torch_device == 'mps' , 'MPS result is not consistent' ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [0.79_21, -0.09_92, -0.19_62, -0.76_95, -0.42_42, 0.78_04, 0.47_37, 0.27_65, 0.33_38] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Optional[int] = SkipDownBlockaD # noqa F405 __magic_name__ :Dict = """down""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_skip_sample=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = [-0.08_45, -0.20_87, -0.24_65, 0.09_71, 0.19_00, -0.04_84, 0.26_64, 0.41_79, 0.50_69] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Any = AttnSkipDownBlockaD # noqa F405 __magic_name__ :List[Any] = """down""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_skip_sample=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = [0.55_39, 0.16_09, 0.49_24, 0.05_37, -0.19_95, 0.40_50, 0.09_79, -0.27_21, -0.06_42] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :List[Any] = DownEncoderBlockaD # noqa F405 __magic_name__ :Dict = """down""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_temb=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Dict = { 'in_channels': 3_2, 'out_channels': 3_2, } lowerCAmelCase__ :List[Any] = self.dummy_input return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Any = [1.11_02, 0.53_02, 0.48_72, -0.00_23, -0.80_42, 0.04_83, -0.34_89, -0.56_32, 0.76_26] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :List[str] = AttnDownEncoderBlockaD # noqa F405 __magic_name__ :Dict = """down""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_temb=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = { 'in_channels': 3_2, 'out_channels': 3_2, } lowerCAmelCase__ :int = self.dummy_input return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Dict = [0.89_66, -0.14_86, 0.85_68, 0.81_41, -0.90_46, -0.13_42, -0.09_72, -0.74_17, 0.15_38] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Tuple = UNetMidBlockaD # noqa F405 __magic_name__ :Dict = """mid""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = { 'in_channels': 3_2, 'temb_channels': 1_2_8, } lowerCAmelCase__ :Union[str, Any] = self.dummy_input return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Dict = [-0.10_62, 1.72_48, 0.34_94, 1.45_69, -0.09_10, -1.24_21, -0.99_84, 0.67_36, 1.00_28] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Dict = UNetMidBlockaDCrossAttn # noqa F405 __magic_name__ :Optional[Any] = """mid""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :int = super().prepare_init_args_and_inputs_for_common() lowerCAmelCase__ :Optional[Any] = 3_2 return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [0.01_87, 2.42_20, 0.44_84, 1.12_03, -0.61_21, -1.51_22, -0.82_70, 0.78_51, 1.83_35] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Optional[Any] = UNetMidBlockaDSimpleCrossAttn # noqa F405 __magic_name__ :Optional[int] = """mid""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_encoder_hidden_states=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :Any = super().prepare_init_args_and_inputs_for_common() lowerCAmelCase__ :Any = 3_2 return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = [0.71_43, 1.99_74, 0.54_48, 1.39_77, 0.12_82, -1.12_37, -1.42_38, 0.55_30, 0.88_80] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Any = UpBlockaD # noqa F405 __magic_name__ :int = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [-0.20_41, -0.41_65, -0.30_22, 0.00_41, -0.66_28, -0.70_53, 0.19_28, -0.03_25, 0.05_23] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :int = ResnetUpsampleBlockaD # noqa F405 __magic_name__ :Any = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [0.22_87, 0.35_49, -0.13_46, 0.47_97, -0.17_15, -0.96_49, 0.73_05, -0.58_64, -0.62_44] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Union[str, Any] = CrossAttnUpBlockaD # noqa F405 __magic_name__ :Union[str, Any] = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :Union[str, Any] = super().prepare_init_args_and_inputs_for_common() lowerCAmelCase__ :Optional[Any] = 3_2 return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = [-0.14_03, -0.35_15, -0.04_20, -0.14_25, 0.31_67, 0.50_94, -0.21_81, 0.59_31, 0.55_82] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Any = SimpleCrossAttnUpBlockaD # noqa F405 __magic_name__ :Union[str, Any] = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase , include_encoder_hidden_states=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :Dict = super().prepare_init_args_and_inputs_for_common() lowerCAmelCase__ :str = 3_2 return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [0.26_45, 0.14_80, 0.09_09, 0.80_44, -0.97_58, -0.90_83, 0.09_94, -1.14_53, -0.74_02] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :List[Any] = AttnUpBlockaD # noqa F405 __magic_name__ :int = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase ) @unittest.skipIf(torch_device == 'mps' , 'MPS result is not consistent' ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = [0.09_79, 0.13_26, 0.00_21, 0.06_59, 0.22_49, 0.00_59, 0.11_32, 0.59_52, 0.10_33] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Any = SkipUpBlockaD # noqa F405 __magic_name__ :int = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = [-0.08_93, -0.12_34, -0.15_06, -0.03_32, 0.01_23, -0.02_11, 0.05_66, 0.01_43, 0.03_62] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :List[str] = AttnSkipUpBlockaD # noqa F405 __magic_name__ :str = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = [0.03_61, 0.06_17, 0.27_87, -0.03_50, 0.03_42, 0.34_21, -0.08_43, 0.09_13, 0.30_15] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :Tuple = UpDecoderBlockaD # noqa F405 __magic_name__ :Union[str, Any] = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_temb=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Any = {'in_channels': 3_2, 'out_channels': 3_2} lowerCAmelCase__ :Dict = self.dummy_input return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = [0.44_04, 0.19_98, -0.98_86, -0.33_20, -0.31_28, -0.70_34, -0.69_55, -0.23_38, -0.31_37] super().test_output(__UpperCAmelCase ) class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :List[str] = AttnUpDecoderBlockaD # noqa F405 __magic_name__ :List[str] = """up""" @property def snake_case ( self ): '''simple docstring''' return super().get_dummy_input(include_temb=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Any = {'in_channels': 3_2, 'out_channels': 3_2} lowerCAmelCase__ :Union[str, Any] = self.dummy_input return init_dict, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = [0.67_38, 0.44_91, 0.10_55, 1.07_10, 0.73_16, 0.33_39, 0.33_52, 0.10_23, 0.35_68] super().test_output(__UpperCAmelCase )
254
"""simple docstring""" import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _lowerCAmelCase ( a ): """simple docstring""" def __init__( self , __UpperCAmelCase = "▁" , __UpperCAmelCase = True , __UpperCAmelCase = "<unk>" , __UpperCAmelCase = "</s>" , __UpperCAmelCase = "<pad>" , ): '''simple docstring''' lowerCAmelCase__ :Tuple = { 'pad': {'id': 0, 'token': pad_token}, 'eos': {'id': 1, 'token': eos_token}, 'unk': {'id': 2, 'token': unk_token}, } lowerCAmelCase__ :Optional[int] = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowerCAmelCase__ :Any = token_dict['token'] lowerCAmelCase__ :int = Tokenizer(Unigram() ) lowerCAmelCase__ :Tuple = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(' {2,}' ) , ' ' ), normalizers.Lowercase(), ] ) lowerCAmelCase__ :Any = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__UpperCAmelCase , add_prefix_space=__UpperCAmelCase ), pre_tokenizers.Digits(individual_digits=__UpperCAmelCase ), pre_tokenizers.Punctuation(), ] ) lowerCAmelCase__ :List[str] = decoders.Metaspace(replacement=__UpperCAmelCase , add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ :Tuple = TemplateProcessing( single=F"$A {self.special_tokens['eos']['token']}" , special_tokens=[(self.special_tokens['eos']['token'], self.special_tokens['eos']['id'])] , ) lowerCAmelCase__ :Optional[int] = { 'model': 'SentencePieceUnigram', 'replacement': replacement, 'add_prefix_space': add_prefix_space, } super().__init__(__UpperCAmelCase , __UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = 8_0_0_0 , __UpperCAmelCase = True , ): '''simple docstring''' lowerCAmelCase__ :int = trainers.UnigramTrainer( vocab_size=__UpperCAmelCase , special_tokens=self.special_tokens_list , show_progress=__UpperCAmelCase , ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ :int = [files] self._tokenizer.train(__UpperCAmelCase , trainer=__UpperCAmelCase ) self.add_unk_id() def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = 8_0_0_0 , __UpperCAmelCase = True , ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = trainers.UnigramTrainer( vocab_size=__UpperCAmelCase , special_tokens=self.special_tokens_list , show_progress=__UpperCAmelCase , ) self._tokenizer.train_from_iterator(__UpperCAmelCase , trainer=__UpperCAmelCase ) self.add_unk_id() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = json.loads(self._tokenizer.to_str() ) lowerCAmelCase__ :List[str] = self.special_tokens['unk']['id'] lowerCAmelCase__ :Union[str, Any] = Tokenizer.from_str(json.dumps(__UpperCAmelCase ) )
254
1
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin __lowercase = get_tests_dir("""fixtures/test_sentencepiece_bpe_char.model""") @require_sentencepiece @require_tokenizers class _A ( _a ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : str = SpeechTaTokenizer UpperCAmelCase : List[Any] = False UpperCAmelCase : Any = True def __snake_case ( self : Optional[Any]): super().setUp() # We have a SentencePiece fixture for testing a : Optional[int] = SpeechTaTokenizer(__UpperCAmelCase) a : List[str] = AddedToken("<mask>" , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase) a : List[Any] = mask_token tokenizer.add_special_tokens({"mask_token": mask_token}) tokenizer.add_tokens(["<ctc_blank>"]) tokenizer.save_pretrained(self.tmpdirname) def __snake_case ( self : Optional[Any] , __UpperCAmelCase : str): a : Dict = "this is a test" a : Tuple = "this is a test" return input_text, output_text def __snake_case ( self : str , __UpperCAmelCase : Dict , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=20 , __UpperCAmelCase : Tuple=5): a , a : List[Any] = self.get_input_output_texts(__UpperCAmelCase) a : Optional[Any] = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase) a : List[str] = tokenizer.decode(__UpperCAmelCase , clean_up_tokenization_spaces=__UpperCAmelCase) return text, ids def __snake_case ( self : Optional[Any]): a : Dict = "<pad>" a : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase) , __UpperCAmelCase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase) , __UpperCAmelCase) def __snake_case ( self : List[Any]): a : Tuple = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , "<s>") self.assertEqual(vocab_keys[1] , "<pad>") self.assertEqual(vocab_keys[-4] , "œ") self.assertEqual(vocab_keys[-2] , "<mask>") self.assertEqual(vocab_keys[-1] , "<ctc_blank>") self.assertEqual(len(__UpperCAmelCase) , 81) def __snake_case ( self : Optional[int]): self.assertEqual(self.get_tokenizer().vocab_size , 79) def __snake_case ( self : int): a : Any = self.get_tokenizers(do_lower_case=__UpperCAmelCase) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}'''): a : int = tokenizer.vocab_size a : Optional[Any] = len(__UpperCAmelCase) self.assertNotEqual(__UpperCAmelCase , 0) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) a : List[str] = ["aaaaa bbbbbb", "cccccccccdddddddd"] a : Tuple = tokenizer.add_tokens(__UpperCAmelCase) a : Optional[int] = tokenizer.vocab_size a : Any = len(__UpperCAmelCase) self.assertNotEqual(__UpperCAmelCase , 0) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase) self.assertEqual(__UpperCAmelCase , len(__UpperCAmelCase)) self.assertEqual(__UpperCAmelCase , all_size + len(__UpperCAmelCase)) a : Union[str, Any] = tokenizer.encode("aaaaa bbbbbb low cccccccccdddddddd l" , add_special_tokens=__UpperCAmelCase) self.assertGreaterEqual(len(__UpperCAmelCase) , 4) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1) a : List[Any] = {"eos_token": ">>>>|||<||<<|<<", "pad_token": "<<<<<|||>|>>>>|>"} a : Any = tokenizer.add_special_tokens(__UpperCAmelCase) a : List[str] = tokenizer.vocab_size a : List[str] = len(__UpperCAmelCase) self.assertNotEqual(__UpperCAmelCase , 0) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase) self.assertEqual(__UpperCAmelCase , len(__UpperCAmelCase)) self.assertEqual(__UpperCAmelCase , all_size_a + len(__UpperCAmelCase)) a : Union[str, Any] = tokenizer.encode( ">>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l" , add_special_tokens=__UpperCAmelCase) self.assertGreaterEqual(len(__UpperCAmelCase) , 6) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1) self.assertGreater(tokens[0] , tokens[1]) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1) self.assertGreater(tokens[-3] , tokens[-4]) self.assertEqual(tokens[0] , tokenizer.eos_token_id) self.assertEqual(tokens[-3] , tokenizer.pad_token_id) def __snake_case ( self : Dict): pass def __snake_case ( self : str): pass def __snake_case ( self : Union[str, Any]): a : Union[str, Any] = self.get_tokenizer() a : List[str] = tokenizer.tokenize("This is a test") # fmt: off self.assertListEqual(__UpperCAmelCase , [SPIECE_UNDERLINE, "T", "h", "i", "s", SPIECE_UNDERLINE, "i", "s", SPIECE_UNDERLINE, "a", SPIECE_UNDERLINE, "t", "e", "s", "t"]) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) a : Any = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( __UpperCAmelCase , [SPIECE_UNDERLINE, "I", SPIECE_UNDERLINE, "w", "a", "s", SPIECE_UNDERLINE, "b", "o", "r", "n", SPIECE_UNDERLINE, "i", "n", SPIECE_UNDERLINE, "92000", ",", SPIECE_UNDERLINE, "a", "n", "d", SPIECE_UNDERLINE, "t", "h", "i", "s", SPIECE_UNDERLINE, "i", "s", SPIECE_UNDERLINE, "f", "a", "l", "s", "é", "."]) a : Tuple = tokenizer.convert_tokens_to_ids(__UpperCAmelCase) # fmt: off self.assertListEqual(__UpperCAmelCase , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26]) # fmt: on a : Optional[int] = tokenizer.convert_ids_to_tokens(__UpperCAmelCase) self.assertListEqual( __UpperCAmelCase , [SPIECE_UNDERLINE, "I", SPIECE_UNDERLINE, "w", "a", "s", SPIECE_UNDERLINE, "b", "o", "r", "n", SPIECE_UNDERLINE, "i", "n", SPIECE_UNDERLINE, "<unk>", ",", SPIECE_UNDERLINE, "a", "n", "d", SPIECE_UNDERLINE, "t", "h", "i", "s", SPIECE_UNDERLINE, "i", "s", SPIECE_UNDERLINE, "f", "a", "l", "s", "é", "."]) @slow def __snake_case ( self : Any): # Use custom sequence because this tokenizer does not handle numbers. a : Union[str, Any] = [ "Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides " "general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural " "Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained " "models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.", "BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly " "conditioning on both left and right context in all layers.", "The quick brown fox jumps over the lazy dog.", ] # fmt: off a : str = { "input_ids": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], "attention_mask": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name="microsoft/speecht5_asr" , revision="c5ef64c71905caeccde0e4462ef3f9077224c524" , sequences=__UpperCAmelCase , )
40
def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = [[] for _ in range(SCREAMING_SNAKE_CASE )] __UpperCAmelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1 or len(SCREAMING_SNAKE_CASE ) <= key: return input_string for position, character in enumerate(SCREAMING_SNAKE_CASE ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = [''''''.join(SCREAMING_SNAKE_CASE ) for row in temp_grid] __UpperCAmelCase = ''''''.join(SCREAMING_SNAKE_CASE ) return output_string def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = [] __UpperCAmelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1: return input_string __UpperCAmelCase = [[] for _ in range(SCREAMING_SNAKE_CASE )] # generates template for position in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append('''*''' ) __UpperCAmelCase = 0 for row in temp_grid: # fills in the characters __UpperCAmelCase = input_string[counter : counter + len(SCREAMING_SNAKE_CASE )] grid.append(list(SCREAMING_SNAKE_CASE ) ) counter += len(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = '''''' # reads as zigzag for position in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def __a ( SCREAMING_SNAKE_CASE ) -> dict[int, str]: '''simple docstring''' __UpperCAmelCase = {} for key_guess in range(1 , len(SCREAMING_SNAKE_CASE ) ): # tries every key __UpperCAmelCase = decrypt(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": import doctest doctest.testmod()
333
0
"""simple docstring""" from __future__ import annotations import requests def lowerCAmelCase (__UpperCamelCase : str ): """simple docstring""" __UpperCamelCase =F"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(__UpperCamelCase ).json() def lowerCAmelCase (__UpperCamelCase : int = 1_0 ): """simple docstring""" __UpperCamelCase ='''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' __UpperCamelCase =requests.get(__UpperCamelCase ).json()[:max_stories] return [get_hackernews_story(__UpperCamelCase ) for story_id in story_ids] def lowerCAmelCase (__UpperCamelCase : int = 1_0 ): """simple docstring""" __UpperCamelCase =hackernews_top_stories(__UpperCamelCase ) return "\n".join('''* [{title}]({url})'''.format(**__UpperCamelCase ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
85
"""simple docstring""" def lowerCAmelCase (__UpperCamelCase : int = 3 , __UpperCamelCase : int = 7 , __UpperCamelCase : int = 1_0_0_0_0_0_0 ): """simple docstring""" __UpperCamelCase =0 __UpperCamelCase =1 for current_denominator in range(1 , limit + 1 ): __UpperCamelCase =current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: __UpperCamelCase =current_numerator __UpperCamelCase =current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_000_000))
85
1
import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["image_processor", "tokenizer"] UpperCamelCase__ = "BlipImageProcessor" UpperCamelCase__ = "AutoTokenizer" def __init__( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" super().__init__(UpperCAmelCase , UpperCAmelCase ) # add QFormer tokenizer _UpperCAmelCase = qformer_tokenizer def __call__( self , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = 0 , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = True , UpperCAmelCase = None , **UpperCAmelCase , ): """simple docstring""" if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) _UpperCAmelCase = BatchFeature() if text is not None: _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) encoding.update(UpperCAmelCase ) _UpperCAmelCase = self.qformer_tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) _UpperCAmelCase = qformer_text_encoding.pop('input_ids' ) _UpperCAmelCase = qformer_text_encoding.pop('attention_mask' ) if images is not None: _UpperCAmelCase = self.image_processor(UpperCAmelCase , return_tensors=UpperCAmelCase ) encoding.update(UpperCAmelCase ) return encoding def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase , **UpperCAmelCase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.tokenizer.model_input_names _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def UpperCamelCase ( self , UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" if os.path.isfile(UpperCAmelCase ): raise ValueError(F"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(UpperCAmelCase ) return super().save_pretrained(UpperCAmelCase , **UpperCAmelCase ) @classmethod def UpperCamelCase ( cls , UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , subfolder='qformer_tokenizer' ) _UpperCAmelCase = cls._get_arguments_from_pretrained(UpperCAmelCase , **UpperCAmelCase ) args.append(UpperCAmelCase ) return cls(*UpperCAmelCase )
39
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class _a ( _lowercase): def UpperCAmelCase__( self : int )-> Union[str, Any]: lowerCAmelCase__ : Optional[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , '''hidden_sizes''' ) ) self.parent.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , '''num_attention_heads''' ) ) self.parent.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , '''num_encoder_blocks''' ) ) class _a : def __init__( self : str , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : Any=13 , _SCREAMING_SNAKE_CASE : List[Any]=64 , _SCREAMING_SNAKE_CASE : str=3 , _SCREAMING_SNAKE_CASE : Union[str, Any]=4 , _SCREAMING_SNAKE_CASE : Optional[int]=[2, 2, 2, 2] , _SCREAMING_SNAKE_CASE : Tuple=[8, 4, 2, 1] , _SCREAMING_SNAKE_CASE : Dict=[16, 32, 64, 128] , _SCREAMING_SNAKE_CASE : Dict=[1, 4, 8, 16] , _SCREAMING_SNAKE_CASE : str=[1, 2, 4, 8] , _SCREAMING_SNAKE_CASE : Dict=True , _SCREAMING_SNAKE_CASE : List[Any]=True , _SCREAMING_SNAKE_CASE : Tuple="gelu" , _SCREAMING_SNAKE_CASE : str=0.1 , _SCREAMING_SNAKE_CASE : List[str]=0.1 , _SCREAMING_SNAKE_CASE : List[Any]=0.02 , _SCREAMING_SNAKE_CASE : Any=3 , _SCREAMING_SNAKE_CASE : Optional[int]=None , )-> List[str]: lowerCAmelCase__ : int = parent lowerCAmelCase__ : Dict = batch_size lowerCAmelCase__ : Dict = image_size lowerCAmelCase__ : Union[str, Any] = num_channels lowerCAmelCase__ : Optional[Any] = num_encoder_blocks lowerCAmelCase__ : Union[str, Any] = sr_ratios lowerCAmelCase__ : int = depths lowerCAmelCase__ : Optional[int] = hidden_sizes lowerCAmelCase__ : Optional[Any] = downsampling_rates lowerCAmelCase__ : Tuple = num_attention_heads lowerCAmelCase__ : Dict = is_training lowerCAmelCase__ : Optional[int] = use_labels lowerCAmelCase__ : Tuple = hidden_act lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : List[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : List[str] = num_labels lowerCAmelCase__ : Union[str, Any] = scope def UpperCAmelCase__( self : Tuple )-> Optional[Any]: lowerCAmelCase__ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[int] = None if self.use_labels: lowerCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCAmelCase__ : Any = self.get_config() return config, pixel_values, labels def UpperCAmelCase__( self : List[str] )-> Optional[int]: return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : Union[str, Any] )-> Any: lowerCAmelCase__ : Union[str, Any] = SegformerModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ : Optional[int] = model(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Any = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def UpperCAmelCase__( self : str , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Optional[Any] )-> Any: lowerCAmelCase__ : Optional[Any] = self.num_labels lowerCAmelCase__ : Tuple = SegformerForSemanticSegmentation(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ : Dict = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) lowerCAmelCase__ : Tuple = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : int )-> Tuple: lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : Tuple = SegformerForSemanticSegmentation(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ : Any = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertGreater(result.loss , 0.0 ) def UpperCAmelCase__( self : Union[str, Any] )-> List[str]: lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : str = config_and_inputs lowerCAmelCase__ : List[Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _a ( _lowercase , _lowercase , unittest.TestCase): _a : Dict = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) _a : Any = ( { '''feature-extraction''': SegformerModel, '''image-classification''': SegformerForImageClassification, '''image-segmentation''': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) _a : List[Any] = True _a : int = False _a : List[str] = False _a : Union[str, Any] = False def UpperCAmelCase__( self : Optional[int] )-> Dict: lowerCAmelCase__ : List[Any] = SegformerModelTester(self ) lowerCAmelCase__ : Optional[Any] = SegformerConfigTester(self , config_class=_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple )-> Optional[int]: self.config_tester.run_common_tests() def UpperCAmelCase__( self : Optional[int] )-> Any: lowerCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Any )-> Dict: lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : List[str] )-> Tuple: lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*_SCREAMING_SNAKE_CASE ) @unittest.skip('''SegFormer does not use inputs_embeds''' ) def UpperCAmelCase__( self : int )-> Dict: pass @unittest.skip('''SegFormer does not have get_input_embeddings method and get_output_embeddings methods''' ) def UpperCAmelCase__( self : str )-> str: pass def UpperCAmelCase__( self : str )-> Any: lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Any = model_class(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : int = [*signature.parameters.keys()] lowerCAmelCase__ : Union[str, Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : List[str] )-> Dict: lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = True for model_class in self.all_model_classes: lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Optional[Any] = False lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : Union[str, Any] = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Tuple = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : Union[str, Any] = outputs.attentions lowerCAmelCase__ : List[str] = sum(self.model_tester.depths ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : int = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Dict = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : str = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # verify the first attentions (first block, first layer) lowerCAmelCase__ : str = (self.model_tester.image_size // 4) ** 2 lowerCAmelCase__ : Optional[int] = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) lowerCAmelCase__ : str = (self.model_tester.image_size // 32) ** 2 lowerCAmelCase__ : Optional[int] = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) lowerCAmelCase__ : int = len(_SCREAMING_SNAKE_CASE ) # Check attention is always last and order is fine lowerCAmelCase__ : Dict = True lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Optional[int] = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) self.assertEqual(out_len + 1 , len(_SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : Optional[int] = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # verify the first attentions (first block, first layer) lowerCAmelCase__ : List[Any] = (self.model_tester.image_size // 4) ** 2 lowerCAmelCase__ : Union[str, Any] = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def UpperCAmelCase__( self : List[str] )-> List[Any]: def check_hidden_states_output(_SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Union[str, Any] ): lowerCAmelCase__ : str = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : Union[str, Any] = outputs.hidden_states lowerCAmelCase__ : Optional[Any] = self.model_tester.num_encoder_blocks self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Dict = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Optional[int] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple )-> Dict: if not self.model_tester.is_training: return lowerCAmelCase__ , lowerCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[Any] = True for model_class in self.all_model_classes: if model_class in get_values(_SCREAMING_SNAKE_CASE ): continue lowerCAmelCase__ : Tuple = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.train() lowerCAmelCase__ : Any = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = model(**_SCREAMING_SNAKE_CASE ).loss loss.backward() @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def UpperCAmelCase__( self : Union[str, Any] )-> Dict: pass @slow def UpperCAmelCase__( self : Union[str, Any] )-> List[Any]: for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Union[str, Any] = SegformerModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( ): """simple docstring""" lowerCAmelCase__ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch class _a ( unittest.TestCase): @slow def UpperCAmelCase__( self : str )-> Any: # only resize + normalize lowerCAmelCase__ : Optional[Any] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_SCREAMING_SNAKE_CASE , align=_SCREAMING_SNAKE_CASE , do_random_crop=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = SegformerForSemanticSegmentation.from_pretrained('''nvidia/segformer-b0-finetuned-ade-512-512''' ).to( _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[int] = prepare_img() lowerCAmelCase__ : Union[str, Any] = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ) lowerCAmelCase__ : Optional[int] = encoded_inputs.pixel_values.to(_SCREAMING_SNAKE_CASE ) with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[Any] = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) ) @slow def UpperCAmelCase__( self : Optional[Any] )-> Any: # only resize + normalize lowerCAmelCase__ : Union[str, Any] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_SCREAMING_SNAKE_CASE , align=_SCREAMING_SNAKE_CASE , do_random_crop=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[int] = SegformerForSemanticSegmentation.from_pretrained( '''nvidia/segformer-b1-finetuned-cityscapes-1024-1024''' ).to(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = prepare_img() lowerCAmelCase__ : Optional[int] = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ) lowerCAmelCase__ : Dict = encoded_inputs.pixel_values.to(_SCREAMING_SNAKE_CASE ) with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1E-1 ) ) @slow def UpperCAmelCase__( self : Any )-> Optional[Any]: # only resize + normalize lowerCAmelCase__ : Dict = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_SCREAMING_SNAKE_CASE , align=_SCREAMING_SNAKE_CASE , do_random_crop=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = SegformerForSemanticSegmentation.from_pretrained('''nvidia/segformer-b0-finetuned-ade-512-512''' ).to( _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[Any] = prepare_img() lowerCAmelCase__ : str = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ) lowerCAmelCase__ : Any = encoded_inputs.pixel_values.to(_SCREAMING_SNAKE_CASE ) with torch.no_grad(): lowerCAmelCase__ : Tuple = model(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Union[str, Any] = outputs.logits.detach().cpu() lowerCAmelCase__ : Union[str, Any] = image_processor.post_process_semantic_segmentation(outputs=_SCREAMING_SNAKE_CASE , target_sizes=[(500, 300)] ) lowerCAmelCase__ : Any = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = image_processor.post_process_semantic_segmentation(outputs=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , _SCREAMING_SNAKE_CASE )
131
0
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class UpperCamelCase__ ( nn.Module ): """simple docstring""" def __init__( self , _A = 16 , _A = 88 , _A = None , _A = 1 , _A = 0.0 , _A = 32 , _A = None , _A = False , _A = None , _A = None , _A = "geglu" , _A = None , ) -> Optional[int]: super().__init__() SCREAMING_SNAKE_CASE_ = nn.ModuleList( [ TransformeraDModel( num_attention_heads=_A , attention_head_dim=_A , in_channels=_A , num_layers=_A , dropout=_A , norm_num_groups=_A , cross_attention_dim=_A , attention_bias=_A , sample_size=_A , num_vector_embeds=_A , activation_fn=_A , num_embeds_ada_norm=_A , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference SCREAMING_SNAKE_CASE_ = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` SCREAMING_SNAKE_CASE_ = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` SCREAMING_SNAKE_CASE_ = [1, 0] def _UpperCamelCase ( self , _A , _A , _A=None , _A=None , _A=None , _A = True , ) -> Any: SCREAMING_SNAKE_CASE_ = hidden_states SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens SCREAMING_SNAKE_CASE_ = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] SCREAMING_SNAKE_CASE_ = self.transformer_index_for_condition[i] SCREAMING_SNAKE_CASE_ = self.transformers[transformer_index]( _A , encoder_hidden_states=_A , timestep=_A , cross_attention_kwargs=_A , return_dict=_A , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] SCREAMING_SNAKE_CASE_ = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) SCREAMING_SNAKE_CASE_ = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=_A )
257
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { "microsoft/unispeech-sat-base-100h-libri-ft": ( "https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json" ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ ="unispeech-sat" def __init__( self , _A=32 , _A=768 , _A=12 , _A=12 , _A=3072 , _A="gelu" , _A=0.1 , _A=0.1 , _A=0.1 , _A=0.0 , _A=0.0 , _A=0.1 , _A=0.1 , _A=0.02 , _A=1E-5 , _A="group" , _A="gelu" , _A=(512, 512, 512, 512, 512, 512, 512) , _A=(5, 2, 2, 2, 2, 2, 2) , _A=(10, 3, 3, 3, 3, 2, 2) , _A=False , _A=128 , _A=16 , _A=False , _A=True , _A=0.05 , _A=10 , _A=2 , _A=0.0 , _A=10 , _A=0 , _A=320 , _A=2 , _A=0.1 , _A=100 , _A=256 , _A=256 , _A=0.1 , _A="mean" , _A=False , _A=False , _A=256 , _A=(512, 512, 512, 512, 1500) , _A=(5, 3, 3, 1, 1) , _A=(1, 2, 3, 1, 1) , _A=512 , _A=0 , _A=1 , _A=2 , _A=504 , **_A , ) -> Tuple: super().__init__(**_A , pad_token_id=_A , bos_token_id=_A , eos_token_id=_A ) SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = feat_extract_norm SCREAMING_SNAKE_CASE_ = feat_extract_activation SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = conv_bias SCREAMING_SNAKE_CASE_ = num_conv_pos_embeddings SCREAMING_SNAKE_CASE_ = num_conv_pos_embedding_groups SCREAMING_SNAKE_CASE_ = len(self.conv_dim ) SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = hidden_dropout SCREAMING_SNAKE_CASE_ = attention_dropout SCREAMING_SNAKE_CASE_ = activation_dropout SCREAMING_SNAKE_CASE_ = feat_proj_dropout SCREAMING_SNAKE_CASE_ = final_dropout SCREAMING_SNAKE_CASE_ = layerdrop SCREAMING_SNAKE_CASE_ = layer_norm_eps SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = num_clusters SCREAMING_SNAKE_CASE_ = do_stable_layer_norm SCREAMING_SNAKE_CASE_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 SCREAMING_SNAKE_CASE_ = apply_spec_augment SCREAMING_SNAKE_CASE_ = mask_time_prob SCREAMING_SNAKE_CASE_ = mask_time_length SCREAMING_SNAKE_CASE_ = mask_time_min_masks SCREAMING_SNAKE_CASE_ = mask_feature_prob SCREAMING_SNAKE_CASE_ = mask_feature_length SCREAMING_SNAKE_CASE_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations SCREAMING_SNAKE_CASE_ = num_codevectors_per_group SCREAMING_SNAKE_CASE_ = num_codevector_groups SCREAMING_SNAKE_CASE_ = contrastive_logits_temperature SCREAMING_SNAKE_CASE_ = feat_quantizer_dropout SCREAMING_SNAKE_CASE_ = num_negatives SCREAMING_SNAKE_CASE_ = codevector_dim SCREAMING_SNAKE_CASE_ = proj_codevector_dim SCREAMING_SNAKE_CASE_ = diversity_loss_weight # ctc loss SCREAMING_SNAKE_CASE_ = ctc_loss_reduction SCREAMING_SNAKE_CASE_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = list(_A ) SCREAMING_SNAKE_CASE_ = xvector_output_dim @property def _UpperCamelCase ( self ) -> str: return functools.reduce(operator.mul , self.conv_stride , 1 )
257
1
"""simple docstring""" import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging UpperCamelCase : Any = "\\n\n" UpperCamelCase : Union[str, Any] = "\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n" UpperCamelCase : Optional[int] = "\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to 'cuda' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"]\n >>> results = perplexity.compute(model_id='gpt2',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 78.22\n >>> print(round(results[\"perplexities\"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = datasets.load_dataset(\"wikitext\",\n ... \"wikitext-2-raw-v1\",\n ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!='']\n >>> results = perplexity.compute(model_id='gpt2',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 60.35\n >>> print(round(results[\"perplexities\"][0], 2))\n 81.12\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): def UpperCAmelCase ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 16 , __UpperCAmelCase = True , __UpperCAmelCase=None ): '''simple docstring''' if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": __UpperCamelCase = 'cuda' else: __UpperCamelCase = 'cuda' if torch.cuda.is_available() else 'cpu' __UpperCamelCase = AutoModelForCausalLM.from_pretrained(__lowerCAmelCase ) __UpperCamelCase = model.to(__lowerCAmelCase ) __UpperCamelCase = AutoTokenizer.from_pretrained(__lowerCAmelCase ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: __UpperCamelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(__lowerCAmelCase ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" __UpperCamelCase = model.config.max_length - 1 else: __UpperCamelCase = model.config.max_length __UpperCamelCase = tokenizer( __lowerCAmelCase , add_special_tokens=__lowerCAmelCase , padding=__lowerCAmelCase , truncation=__lowerCAmelCase , max_length=__lowerCAmelCase , return_tensors='pt' , return_attention_mask=__lowerCAmelCase , ).to(__lowerCAmelCase ) __UpperCamelCase = encodings['input_ids'] __UpperCamelCase = encodings['attention_mask'] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." __UpperCamelCase = [] __UpperCamelCase = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(__lowerCAmelCase ) , __lowerCAmelCase ) ): __UpperCamelCase = min(start_index + batch_size , len(__lowerCAmelCase ) ) __UpperCamelCase = encoded_texts[start_index:end_index] __UpperCamelCase = attn_masks[start_index:end_index] if add_start_token: __UpperCamelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(__lowerCAmelCase ) __UpperCamelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) __UpperCamelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(__lowerCAmelCase ), attn_mask] , dim=1 ) __UpperCamelCase = encoded_batch with torch.no_grad(): __UpperCamelCase = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase ).logits __UpperCamelCase = out_logits[..., :-1, :].contiguous() __UpperCamelCase = labels[..., 1:].contiguous() __UpperCamelCase = attn_mask[..., 1:].contiguous() __UpperCamelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , __lowerCAmelCase ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(__lowerCAmelCase )}
316
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = "▁" _a = { "vocab_file": "vocab.json", "spm_file": "sentencepiece.bpe.model", "tokenizer_config_file": "tokenizer_config.json", } _a = { "vocab_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json", }, "spm_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model", }, "tokenizer_config_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json", }, } _a = { "facebook/m2m100_418M": 1_024, } # fmt: off _a = { "m2m100": ["af", "am", "ar", "ast", "az", "ba", "be", "bg", "bn", "br", "bs", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "fa", "ff", "fi", "fr", "fy", "ga", "gd", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hy", "id", "ig", "ilo", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "lb", "lg", "ln", "lo", "lt", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no", "ns", "oc", "or", "pa", "pl", "ps", "pt", "ro", "ru", "sd", "si", "sk", "sl", "so", "sq", "sr", "ss", "su", "sv", "sw", "ta", "th", "tl", "tn", "tr", "uk", "ur", "uz", "vi", "wo", "xh", "yi", "yo", "zh", "zu"], "wmt21": ["en", "ha", "is", "ja", "cs", "ru", "zh", "de"] } class __A ( lowerCAmelCase ): '''simple docstring''' lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = ["""input_ids""", """attention_mask"""] lowerCAmelCase_ = [] lowerCAmelCase_ = [] def __init__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase="<s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="<pad>" , __lowerCAmelCase="<unk>" , __lowerCAmelCase="m2m100" , __lowerCAmelCase = None , __lowerCAmelCase=8 , **__lowerCAmelCase , ): '''simple docstring''' lowerCamelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs lowerCamelCase__ = language_codes lowerCamelCase__ = FAIRSEQ_LANGUAGE_CODES[language_codes] lowerCamelCase__ = {lang_code: F'__{lang_code}__' for lang_code in fairseq_language_code} lowerCamelCase__ = kwargs.get('''additional_special_tokens''' , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(__lowerCAmelCase ) for lang_code in fairseq_language_code if self.get_lang_token(__lowerCAmelCase ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__lowerCAmelCase , tgt_lang=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , sep_token=__lowerCAmelCase , unk_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , language_codes=__lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=__lowerCAmelCase , **__lowerCAmelCase , ) lowerCamelCase__ = vocab_file lowerCamelCase__ = load_json(__lowerCAmelCase ) lowerCamelCase__ = {v: k for k, v in self.encoder.items()} lowerCamelCase__ = spm_file lowerCamelCase__ = load_spm(__lowerCAmelCase , self.sp_model_kwargs ) lowerCamelCase__ = len(self.encoder ) lowerCamelCase__ = { self.get_lang_token(__lowerCAmelCase ): self.encoder_size + i for i, lang_code in enumerate(__lowerCAmelCase ) } lowerCamelCase__ = {lang_code: self.encoder_size + i for i, lang_code in enumerate(__lowerCAmelCase )} lowerCamelCase__ = {v: k for k, v in self.lang_token_to_id.items()} lowerCamelCase__ = src_lang if src_lang is not None else '''en''' lowerCamelCase__ = tgt_lang lowerCamelCase__ = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) lowerCamelCase__ = num_madeup_words @property def __lowerCamelCase ( self ): '''simple docstring''' return len(self.encoder ) + len(self.lang_token_to_id ) @property def __lowerCamelCase ( self ): '''simple docstring''' return self._src_lang @src_lang.setter def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' return self.sp_model.encode(__lowerCAmelCase , out_type=__lowerCAmelCase ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(__lowerCAmelCase , self.encoder[self.unk_token] ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(__lowerCAmelCase , self.unk_token ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = [] lowerCamelCase__ = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__lowerCAmelCase ) + token lowerCamelCase__ = [] else: current_sub_tokens.append(__lowerCAmelCase ) out_string += self.sp_model.decode(__lowerCAmelCase ) return out_string.strip() def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCAmelCase , token_ids_a=__lowerCAmelCase , already_has_special_tokens=__lowerCAmelCase ) lowerCamelCase__ = [1] * len(self.prefix_tokens ) lowerCamelCase__ = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__lowerCAmelCase )) + suffix_ones return prefix_ones + ([0] * len(__lowerCAmelCase )) + ([0] * len(__lowerCAmelCase )) + suffix_ones def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowerCamelCase ( self ): '''simple docstring''' lowerCamelCase__ = {self.convert_ids_to_tokens(__lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' lowerCamelCase__ = self.__dict__.copy() lowerCamelCase__ = None return state def __setstate__( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowerCamelCase__ = {} lowerCamelCase__ = load_spm(self.spm_file , self.sp_model_kwargs ) def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): '''simple docstring''' lowerCamelCase__ = Path(__lowerCAmelCase ) if not save_dir.is_dir(): raise OSError(F'{save_directory} should be a directory' ) lowerCamelCase__ = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) lowerCamelCase__ = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder , __lowerCAmelCase ) if os.path.abspath(self.spm_file ) != os.path.abspath(__lowerCAmelCase ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , __lowerCAmelCase ) elif not os.path.isfile(self.spm_file ): with open(__lowerCAmelCase , '''wb''' ) as fi: lowerCamelCase__ = self.sp_model.serialized_model_proto() fi.write(__lowerCAmelCase ) return (str(__lowerCAmelCase ), str(__lowerCAmelCase )) def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = "en" , __lowerCAmelCase = None , __lowerCAmelCase = "ro" , **__lowerCAmelCase , ): '''simple docstring''' lowerCamelCase__ = src_lang lowerCamelCase__ = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(__lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ) def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) lowerCamelCase__ = src_lang lowerCamelCase__ = self(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , **__lowerCAmelCase ) lowerCamelCase__ = self.get_lang_id(__lowerCAmelCase ) lowerCamelCase__ = tgt_lang_id return inputs def __lowerCamelCase ( self ): '''simple docstring''' self.set_src_lang_special_tokens(self.src_lang ) def __lowerCamelCase ( self ): '''simple docstring''' self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = self.get_lang_token(__lowerCAmelCase ) lowerCamelCase__ = self.lang_token_to_id[lang_token] lowerCamelCase__ = [self.cur_lang_id] lowerCamelCase__ = [self.eos_token_id] def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = self.get_lang_token(__lowerCAmelCase ) lowerCamelCase__ = self.lang_token_to_id[lang_token] lowerCamelCase__ = [self.cur_lang_id] lowerCamelCase__ = [self.eos_token_id] def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' return self.lang_code_to_token[lang] def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = self.get_lang_token(__lowerCAmelCase ) return self.lang_token_to_id[lang_token] def lowerCAmelCase__(__snake_case ,__snake_case ) -> sentencepiece.SentencePieceProcessor: '''simple docstring''' lowerCamelCase__ = sentencepiece.SentencePieceProcessor(**__snake_case ) spm.Load(str(__snake_case ) ) return spm def lowerCAmelCase__(__snake_case ) -> Union[Dict, List]: '''simple docstring''' with open(__snake_case ,'''r''' ) as f: return json.load(__snake_case ) def lowerCAmelCase__(__snake_case ,__snake_case ) -> None: '''simple docstring''' with open(__snake_case ,'''w''' ) as f: json.dump(__snake_case ,__snake_case ,indent=2 )
209
0
'''simple docstring''' __UpperCAmelCase :dict[tuple[int, int, int], int] = {} def _a ( _lowercase : int , _lowercase : int , _lowercase : int ): '''simple docstring''' if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on __UpperCAmelCase : Tuple = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one __UpperCAmelCase : Dict = _calculate(days - 1 , _lowercase , late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 __UpperCAmelCase : int = _calculate(days - 1 , absent + 1 , 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter __UpperCAmelCase : Tuple = _calculate(days - 1 , _lowercase , 0 ) __UpperCAmelCase : List[str] = state_late + state_absent + state_ontime __UpperCAmelCase : str = prizestrings return prizestrings def _a ( _lowercase : int = 30 ): '''simple docstring''' return _calculate(_lowercase , absent=0 , late=0 ) if __name__ == "__main__": print(solution())
364
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase :Any = logging.get_logger(__name__) __UpperCAmelCase :Dict = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def _a ( _lowercase : Tuple ): '''simple docstring''' if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase : Any = k.replace(_lowercase , _lowercase ) if k.startswith('''encoder''' ): __UpperCAmelCase : str = k.replace('''.attn''' , '''.self_attn''' ) __UpperCAmelCase : Any = k.replace('''norm1''' , '''self_attn_layer_norm''' ) __UpperCAmelCase : List[str] = k.replace('''norm2''' , '''final_layer_norm''' ) elif k.startswith('''decoder''' ): __UpperCAmelCase : int = k.replace('''norm1''' , '''self_attn_layer_norm''' ) __UpperCAmelCase : Union[str, Any] = k.replace('''norm2''' , '''encoder_attn_layer_norm''' ) __UpperCAmelCase : List[Any] = k.replace('''norm3''' , '''final_layer_norm''' ) return k def _a ( _lowercase : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : int = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: __UpperCAmelCase : Any = sd.pop(_lowercase ) __UpperCAmelCase : Optional[int] = k.replace('''layernorm_embedding''' , '''layer_norm''' ) assert new_k not in sd __UpperCAmelCase : List[str] = v __UpperCAmelCase :str = ["START"] @torch.no_grad() def _a ( _lowercase : Optional[int] , _lowercase : Optional[int] , _lowercase : str ): '''simple docstring''' __UpperCAmelCase : Any = torch.load(_lowercase , map_location='''cpu''' ) __UpperCAmelCase : List[str] = model['''model'''] __UpperCAmelCase : Optional[Any] = BlenderbotConfig.from_json_file(_lowercase ) __UpperCAmelCase : Optional[Any] = BlenderbotForConditionalGeneration(_lowercase ) __UpperCAmelCase : Optional[Any] = m.model.state_dict().keys() __UpperCAmelCase : int = [] __UpperCAmelCase : List[str] = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase : int = rename_state_dict_key(_lowercase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase : Union[str, Any] = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(_lowercase ) m.model.load_state_dict(_lowercase , strict=_lowercase ) m.half() m.save_pretrained(_lowercase ) if __name__ == "__main__": __UpperCAmelCase :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) __UpperCAmelCase :Tuple = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
240
0
'''simple docstring''' import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" UpperCAmelCase__ = 0 @slow def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(_UpperCAmelCase ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(_UpperCAmelCase ) , 0 ) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 20 ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" UpperCAmelCase__ = AutoConfig.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) # Check that tokenizer_type ≠ model_type UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , config=_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.txt""" , os.path.join(_UpperCAmelCase , """vocab.txt""" ) ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , tokenizer_type="""bert""" , use_fast=_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.json""" , os.path.join(_UpperCAmelCase , """vocab.json""" ) ) shutil.copy("""./tests/fixtures/merges.txt""" , os.path.join(_UpperCAmelCase , """merges.txt""" ) ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , tokenizer_type="""gpt2""" , use_fast=_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.txt""" , os.path.join(_UpperCAmelCase , """vocab.txt""" ) ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , tokenizer_type="""bert""" ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.json""" , os.path.join(_UpperCAmelCase , """vocab.json""" ) ) shutil.copy("""./tests/fixtures/merges.txt""" , os.path.join(_UpperCAmelCase , """merges.txt""" ) ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , tokenizer_type="""gpt2""" ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" with pytest.raises(_UpperCAmelCase ): AutoTokenizer.from_pretrained("""./""" , tokenizer_type="""xxx""" ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: UpperCAmelCase__ = tokenizer_class.from_pretrained("""wietsedv/bert-base-dutch-cased""" ) self.assertIsInstance(_UpperCAmelCase , (BertTokenizer, BertTokenizerFast) ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , _UpperCAmelCase ) else: self.assertEqual(tokenizer.do_lower_case , _UpperCAmelCase ) self.assertEqual(tokenizer.model_max_length , 5_12 ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( _UpperCAmelCase , """julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier""" , ): UpperCAmelCase__ = tokenizer_class.from_pretrained("""julien-c/herlolip-not-exists""" ) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" UpperCAmelCase__ = TOKENIZER_MAPPING.values() UpperCAmelCase__ = [] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(_UpperCAmelCase ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" self.assertIsInstance(AutoTokenizer.from_pretrained("""bert-base-cased""" , use_fast=_UpperCAmelCase ) , _UpperCAmelCase ) self.assertIsInstance(AutoTokenizer.from_pretrained("""bert-base-cased""" ) , _UpperCAmelCase ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained("""distilbert-base-uncased""" , do_lower_case=_UpperCAmelCase ) UpperCAmelCase__ = """Hello, world. How are you?""" UpperCAmelCase__ = tokenizer.tokenize(_UpperCAmelCase ) self.assertEqual("""[UNK]""" , tokens[0] ) UpperCAmelCase__ = AutoTokenizer.from_pretrained("""microsoft/mpnet-base""" , do_lower_case=_UpperCAmelCase ) UpperCAmelCase__ = tokenizer.tokenize(_UpperCAmelCase ) self.assertEqual("""[UNK]""" , tokens[0] ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained("""robot-test/dummy-tokenizer-fast-with-model-config""" ) self.assertEqual(type(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(tokenizer.model_max_length , 5_12 ) self.assertEqual(tokenizer.vocab_size , 3_00_00 ) self.assertEqual(tokenizer.unk_token , """[UNK]""" ) self.assertEqual(tokenizer.padding_side , """right""" ) self.assertEqual(tokenizer.truncation_side , """right""" ) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 12 ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained("""ctrl""" ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" UpperCAmelCase__ = get_tokenizer_config("""bert-base-cased""" ) UpperCAmelCase__ = config.pop("""_commit_hash""" , _UpperCAmelCase ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(_UpperCAmelCase , {"""do_lower_case""": False} ) # This model does not have a tokenizer_config so we get back an empty dict. UpperCAmelCase__ = get_tokenizer_config(_UpperCAmelCase ) self.assertDictEqual(_UpperCAmelCase , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = get_tokenizer_config(_UpperCAmelCase ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config["""tokenizer_class"""] , """BertTokenizer""" ) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" try: AutoConfig.register("""custom""" , _UpperCAmelCase ) AutoTokenizer.register(_UpperCAmelCase , slow_tokenizer_class=_UpperCAmelCase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_UpperCAmelCase ): AutoTokenizer.register(_UpperCAmelCase , slow_tokenizer_class=_UpperCAmelCase ) UpperCAmelCase__ = CustomTokenizer.from_pretrained(_UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" try: AutoConfig.register("""custom""" , _UpperCAmelCase ) # Can register in two steps AutoTokenizer.register(_UpperCAmelCase , slow_tokenizer_class=_UpperCAmelCase ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(_UpperCAmelCase , fast_tokenizer_class=_UpperCAmelCase ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( _UpperCAmelCase , slow_tokenizer_class=_UpperCAmelCase , fast_tokenizer_class=_UpperCAmelCase ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_UpperCAmelCase ): AutoTokenizer.register(_UpperCAmelCase , fast_tokenizer_class=_UpperCAmelCase ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase__ = BertTokenizerFast.from_pretrained(_UpperCAmelCase ) bert_tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = CustomTokenizerFast.from_pretrained(_UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , use_fast=_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" with self.assertRaises(_UpperCAmelCase ): UpperCAmelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(_UpperCAmelCase ): UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , trust_remote_code=_UpperCAmelCase ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase , use_fast=_UpperCAmelCase ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , trust_remote_code=_UpperCAmelCase , use_fast=_UpperCAmelCase ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizer""" ) @require_tokenizers def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : Dict = False class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : int = NewTokenizer lowerCAmelCase_ : Any = False try: AutoConfig.register("""custom""" , _UpperCAmelCase ) AutoTokenizer.register(_UpperCAmelCase , slow_tokenizer_class=_UpperCAmelCase ) AutoTokenizer.register(_UpperCAmelCase , fast_tokenizer_class=_UpperCAmelCase ) # If remote code is not set, the default is to use local UpperCAmelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertFalse(tokenizer.special_attribute_present ) UpperCAmelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" , use_fast=_UpperCAmelCase ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertFalse(tokenizer.special_attribute_present ) UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase , use_fast=_UpperCAmelCase ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertTrue(tokenizer.special_attribute_present ) UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=_UpperCAmelCase , use_fast=_UpperCAmelCase ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer_legacy""" , trust_remote_code=_UpperCAmelCase ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version UpperCAmelCase__ = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer_legacy""" , trust_remote_code=_UpperCAmelCase , use_fast=_UpperCAmelCase ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" with self.assertRaisesRegex( _UpperCAmelCase , """bert-base is not a local folder and is not a valid model identifier""" ): UpperCAmelCase__ = AutoTokenizer.from_pretrained("""bert-base""" ) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" with self.assertRaisesRegex( _UpperCAmelCase , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase , revision="""aaaaaa""" ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" UpperCAmelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) with RequestCounter() as counter: UpperCAmelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
346
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging UpperCAmelCase_ = '\\n\n' UpperCAmelCase_ = '\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n' UpperCAmelCase_ = '\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to \'cuda\' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric("perplexity")\n >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"]\n >>> results = perplexity.compute(model_id=\'gpt2\',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n [\'perplexities\', \'mean_perplexity\']\n >>> print(round(results["mean_perplexity"], 2))\n 78.22\n >>> print(round(results["perplexities"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric("perplexity")\n >>> input_texts = datasets.load_dataset("wikitext",\n ... "wikitext-2-raw-v1",\n ... split="test")["text"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!=\'\']\n >>> results = perplexity.compute(model_id=\'gpt2\',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n [\'perplexities\', \'mean_perplexity\']\n >>> print(round(results["mean_perplexity"], 2))\n 60.35\n >>> print(round(results["perplexities"][0], 2))\n 81.12\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_ ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """input_texts""": datasets.Value("""string""" ), } ) , reference_urls=["""https://huggingface.co/docs/transformers/perplexity"""] , ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : int , _UpperCAmelCase : int = 16 , _UpperCAmelCase : bool = True , _UpperCAmelCase : Optional[int]=None ): """simple docstring""" if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCAmelCase__ = """cuda""" else: UpperCAmelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" UpperCAmelCase__ = AutoModelForCausalLM.from_pretrained(_UpperCAmelCase ) UpperCAmelCase__ = model.to(_UpperCAmelCase ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(_UpperCAmelCase ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCAmelCase__ = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(_UpperCAmelCase ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"""pad_token""": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCAmelCase__ = model.config.max_length - 1 else: UpperCAmelCase__ = model.config.max_length UpperCAmelCase__ = tokenizer( _UpperCAmelCase , add_special_tokens=_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase , return_tensors="""pt""" , return_attention_mask=_UpperCAmelCase , ).to(_UpperCAmelCase ) UpperCAmelCase__ = encodings["""input_ids"""] UpperCAmelCase__ = encodings["""attention_mask"""] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCAmelCase__ = [] UpperCAmelCase__ = CrossEntropyLoss(reduction="""none""" ) for start_index in logging.tqdm(range(0 , len(_UpperCAmelCase ) , _UpperCAmelCase ) ): UpperCAmelCase__ = min(start_index + batch_size , len(_UpperCAmelCase ) ) UpperCAmelCase__ = encoded_texts[start_index:end_index] UpperCAmelCase__ = attn_masks[start_index:end_index] if add_start_token: UpperCAmelCase__ = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(_UpperCAmelCase ) UpperCAmelCase__ = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCAmelCase__ = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(_UpperCAmelCase ), attn_mask] , dim=1 ) UpperCAmelCase__ = encoded_batch with torch.no_grad(): UpperCAmelCase__ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase ).logits UpperCAmelCase__ = out_logits[..., :-1, :].contiguous() UpperCAmelCase__ = labels[..., 1:].contiguous() UpperCAmelCase__ = attn_mask[..., 1:].contiguous() UpperCAmelCase__ = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , _UpperCAmelCase ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(_UpperCAmelCase )}
346
1
"""simple docstring""" import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def _a ( _SCREAMING_SNAKE_CASE = 8 ) -> str: snake_case_ = ascii_letters + digits + punctuation return "".join(secrets.choice(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ) ) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: # Password Generator = full boot with random_number, random_letters, and # random_character FUNCTIONS # Put your code here... i -= len(_SCREAMING_SNAKE_CASE ) snake_case_ = i // 3 snake_case_ = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) snake_case_ = ( chars_incl + random(_SCREAMING_SNAKE_CASE , quotient + remainder ) + random(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) + random(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) snake_case_ = list(_SCREAMING_SNAKE_CASE ) shuffle(_SCREAMING_SNAKE_CASE ) return "".join(_SCREAMING_SNAKE_CASE ) # random is a generalised function for letters, characters and numbers def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: return "".join(secrets.choice(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ) ) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Any: pass # Put your code here... def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Optional[int]: pass # Put your code here... def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: pass # Put your code here... def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 8 ) -> bool: if len(_SCREAMING_SNAKE_CASE ) < min_length: # Your Password must be at least 8 characters long return False snake_case_ = any(char in ascii_uppercase for char in password ) snake_case_ = any(char in ascii_lowercase for char in password ) snake_case_ = any(char in digits for char in password ) snake_case_ = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def _a ( ) -> Union[str, Any]: snake_case_ = int(input("""Please indicate the max length of your password: """ ).strip() ) snake_case_ = input( """Please indicate the characters that must be in your password: """ ).strip() print("""Password generated:""" , password_generator(_SCREAMING_SNAKE_CASE ) ) print( """Alternative Password generated:""" , alternative_password_generator(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , ) print("""[If you are thinking of using this passsword, You better save it.]""" ) if __name__ == "__main__": main()
357
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class __A (unittest.TestCase): '''simple docstring''' def lowerCAmelCase ( self : Dict ) ->List[Any]: """simple docstring""" snake_case_ = """laion/clap-htsat-unfused""" snake_case_ = tempfile.mkdtemp() def lowerCAmelCase ( self : List[str] , **UpperCAmelCase_ : Tuple ) ->str: """simple docstring""" return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase_ ) def lowerCAmelCase ( self : Optional[Any] , **UpperCAmelCase_ : Any ) ->Optional[Any]: """simple docstring""" return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase_ ) def lowerCAmelCase ( self : str ) ->Any: """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCAmelCase ( self : Tuple ) ->Optional[int]: """simple docstring""" snake_case_ = self.get_tokenizer() snake_case_ = self.get_feature_extractor() snake_case_ = ClapProcessor(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ ) processor.save_pretrained(self.tmpdirname ) snake_case_ = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase_ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase_ ) def lowerCAmelCase ( self : int ) ->List[str]: """simple docstring""" snake_case_ = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) snake_case_ = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) snake_case_ = self.get_feature_extractor(do_normalize=UpperCAmelCase_ , padding_value=1.0 ) snake_case_ = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase_ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase_ ) def lowerCAmelCase ( self : str ) ->Dict: """simple docstring""" snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = ClapProcessor(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ ) snake_case_ = floats_list((3, 1_000) ) snake_case_ = feature_extractor(UpperCAmelCase_ , return_tensors="""np""" ) snake_case_ = processor(audios=UpperCAmelCase_ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowerCAmelCase ( self : Union[str, Any] ) ->Optional[int]: """simple docstring""" snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = ClapProcessor(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ ) snake_case_ = """This is a test string""" snake_case_ = processor(text=UpperCAmelCase_ ) snake_case_ = tokenizer(UpperCAmelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCAmelCase ( self : Optional[int] ) ->Tuple: """simple docstring""" snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = ClapProcessor(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ ) snake_case_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] snake_case_ = processor.batch_decode(UpperCAmelCase_ ) snake_case_ = tokenizer.batch_decode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCAmelCase ( self : List[str] ) ->Union[str, Any]: """simple docstring""" snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = ClapProcessor(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
233
0
import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py _a = '''.''' if __name__ == "__main__": _a = os.path.join(REPO_PATH, '''utils/documentation_tests.txt''') _a = [] _a = [] with open(doctest_file_path) as fp: for line in fp: _a = line.strip() _a = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: _a = '''\n'''.join(non_existent_paths) raise ValueError(F'''`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}''') if all_paths != sorted(all_paths): raise ValueError('''Files in `utils/documentation_tests.txt` are not in alphabetical order.''')
39
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class __lowerCAmelCase ( lowerCAmelCase_ ): """simple docstring""" A__ : Any = '''EncodecFeatureExtractor''' A__ : Optional[int] = ('''T5Tokenizer''', '''T5TokenizerFast''') def __init__( self : Optional[int] , _snake_case : Optional[Any] , _snake_case : Optional[Any] ): super().__init__(_snake_case , _snake_case ) __lowercase : List[Any] = self.feature_extractor __lowercase : Tuple = False def snake_case_ ( self : Optional[int] , _snake_case : Union[str, Any]=None , _snake_case : Optional[Any]=None , _snake_case : List[str]=True ): return self.tokenizer.get_decoder_prompt_ids(task=_snake_case , language=_snake_case , no_timestamps=_snake_case ) def __call__( self : str , *_snake_case : Tuple , **_snake_case : str ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_snake_case , **_snake_case ) __lowercase : Optional[Any] = kwargs.pop('''audio''' , _snake_case ) __lowercase : str = kwargs.pop('''sampling_rate''' , _snake_case ) __lowercase : Any = kwargs.pop('''text''' , _snake_case ) if len(_snake_case ) > 0: __lowercase : Dict = args[0] __lowercase : Any = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if text is not None: __lowercase : str = self.tokenizer(_snake_case , **_snake_case ) if audio is not None: __lowercase : List[str] = self.feature_extractor(_snake_case , *_snake_case , sampling_rate=_snake_case , **_snake_case ) if audio is None: return inputs elif text is None: return audio_inputs else: __lowercase : Tuple = audio_inputs['''input_values'''] if "padding_mask" in audio_inputs: __lowercase : Tuple = audio_inputs['''padding_mask'''] return inputs def snake_case_ ( self : int , *_snake_case : int , **_snake_case : Any ): __lowercase : Dict = kwargs.pop('''audio''' , _snake_case ) __lowercase : Tuple = kwargs.pop('''padding_mask''' , _snake_case ) if len(_snake_case ) > 0: __lowercase : str = args[0] __lowercase : Tuple = args[1:] if audio_values is not None: return self._decode_audio(_snake_case , padding_mask=_snake_case ) else: return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def snake_case_ ( self : Optional[int] , *_snake_case : int , **_snake_case : List[str] ): return self.tokenizer.decode(*_snake_case , **_snake_case ) def snake_case_ ( self : Dict , _snake_case : List[Any] , _snake_case : Optional = None ): __lowercase : Union[str, Any] = to_numpy(_snake_case ) __lowercase , __lowercase , __lowercase : Optional[int] = audio_values.shape if padding_mask is None: return list(_snake_case ) __lowercase : Optional[int] = to_numpy(_snake_case ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) __lowercase : int = seq_len - padding_mask.shape[-1] __lowercase : Optional[int] = 1 - self.feature_extractor.padding_value __lowercase : Tuple = np.pad(_snake_case , ((0, 0), (0, difference)) , '''constant''' , constant_values=_snake_case ) __lowercase : str = audio_values.tolist() for i in range(_snake_case ): __lowercase : str = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] __lowercase : Any = sliced_audio.reshape(_snake_case , -1 ) return audio_values
156
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Tuple = { '''transfo-xl-wt103''': '''https://huggingface.co/transfo-xl-wt103/resolve/main/config.json''', } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "transfo-xl" a = ["mems"] a = { "n_token": "vocab_size", "hidden_size": "d_model", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any , __lowerCamelCase : int=26_7735 , __lowerCamelCase : Any=[2_0000, 4_0000, 20_0000] , __lowerCamelCase : Dict=1024 , __lowerCamelCase : Optional[int]=1024 , __lowerCamelCase : List[Any]=16 , __lowerCamelCase : Union[str, Any]=64 , __lowerCamelCase : Dict=4096 , __lowerCamelCase : int=4 , __lowerCamelCase : Dict=False , __lowerCamelCase : Tuple=18 , __lowerCamelCase : Optional[int]=1600 , __lowerCamelCase : str=1000 , __lowerCamelCase : Any=True , __lowerCamelCase : Dict=True , __lowerCamelCase : List[Any]=0 , __lowerCamelCase : int=-1 , __lowerCamelCase : Tuple=True , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : int=0.0 , __lowerCamelCase : int=True , __lowerCamelCase : str="normal" , __lowerCamelCase : List[str]=0.01 , __lowerCamelCase : Any=0.01 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : List[str]=1e-5 , __lowerCamelCase : Union[str, Any]=0 , **__lowerCamelCase : int , ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = [] self.cutoffs.extend(__lowerCamelCase ) if proj_share_all_but_first: SCREAMING_SNAKE_CASE__ = [False] + [True] * len(self.cutoffs ) else: SCREAMING_SNAKE_CASE__ = [False] + [False] * len(self.cutoffs ) SCREAMING_SNAKE_CASE__ = d_model SCREAMING_SNAKE_CASE__ = d_embed SCREAMING_SNAKE_CASE__ = d_head SCREAMING_SNAKE_CASE__ = d_inner SCREAMING_SNAKE_CASE__ = div_val SCREAMING_SNAKE_CASE__ = pre_lnorm SCREAMING_SNAKE_CASE__ = n_layer SCREAMING_SNAKE_CASE__ = n_head SCREAMING_SNAKE_CASE__ = mem_len SCREAMING_SNAKE_CASE__ = same_length SCREAMING_SNAKE_CASE__ = attn_type SCREAMING_SNAKE_CASE__ = clamp_len SCREAMING_SNAKE_CASE__ = sample_softmax SCREAMING_SNAKE_CASE__ = adaptive SCREAMING_SNAKE_CASE__ = dropout SCREAMING_SNAKE_CASE__ = dropatt SCREAMING_SNAKE_CASE__ = untie_r SCREAMING_SNAKE_CASE__ = init SCREAMING_SNAKE_CASE__ = init_range SCREAMING_SNAKE_CASE__ = proj_init_std SCREAMING_SNAKE_CASE__ = init_std SCREAMING_SNAKE_CASE__ = layer_norm_epsilon super().__init__(eos_token_id=__lowerCamelCase , **__lowerCamelCase ) @property def lowercase_ ( self : str ) -> Dict: # Message copied from Transformer-XL documentation logger.info(f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def lowercase_ ( self : List[str] , __lowerCamelCase : Any ) -> List[Any]: # Message copied from Transformer-XL documentation raise NotImplementedError( f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
357
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
218
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyImgaImgPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __magic_name__ ( _UpperCAmelCase, unittest.TestCase): UpperCamelCase__ = KandinskyImgaImgPipeline UpperCamelCase__ = ['''prompt''', '''image_embeds''', '''negative_image_embeds''', '''image'''] UpperCamelCase__ = [ '''prompt''', '''negative_prompt''', '''image_embeds''', '''negative_image_embeds''', '''image''', ] UpperCamelCase__ = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''negative_prompt''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] UpperCamelCase__ = False @property def SCREAMING_SNAKE_CASE_ ( self : Dict ): return 32 @property def SCREAMING_SNAKE_CASE_ ( self : str ): return 32 @property def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): return self.time_input_dim @property def SCREAMING_SNAKE_CASE_ ( self : int ): return self.time_input_dim * 4 @property def SCREAMING_SNAKE_CASE_ ( self : Dict ): return 100 @property def SCREAMING_SNAKE_CASE_ ( self : Dict ): lowercase_ : str = XLMRobertaTokenizerFast.from_pretrained("""YiYiXu/tiny-random-mclip-base""" ) return tokenizer @property def SCREAMING_SNAKE_CASE_ ( self : List[str] ): torch.manual_seed(0 ) lowercase_ : List[str] = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1005 , ) lowercase_ : int = MultilingualCLIP(lowercase_ ) lowercase_ : Union[str, Any] = text_encoder.eval() return text_encoder @property def SCREAMING_SNAKE_CASE_ ( self : Dict ): torch.manual_seed(0 ) lowercase_ : Union[str, Any] = { """in_channels""": 4, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """text_image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """text_image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } lowercase_ : List[str] = UNetaDConditionModel(**lowercase_ ) return model @property def SCREAMING_SNAKE_CASE_ ( self : List[str] ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def SCREAMING_SNAKE_CASE_ ( self : Dict ): torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : Dict = self.dummy_text_encoder lowercase_ : Tuple = self.dummy_tokenizer lowercase_ : str = self.dummy_unet lowercase_ : Tuple = self.dummy_movq lowercase_ : List[Any] = { """num_train_timesteps""": 1000, """beta_schedule""": """linear""", """beta_start""": 0.0_00_85, """beta_end""": 0.0_12, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } lowercase_ : Tuple = DDIMScheduler(**lowercase_ ) lowercase_ : List[str] = { """text_encoder""": text_encoder, """tokenizer""": tokenizer, """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def SCREAMING_SNAKE_CASE_ ( self : Tuple , lowercase_ : Any , lowercase_ : List[str]=0 ): lowercase_ : List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) lowercase_ : Any = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(lowercase_ ) # create init_image lowercase_ : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) lowercase_ : List[str] = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Any = Image.fromarray(np.uinta(lowercase_ ) ).convert("""RGB""" ).resize((256, 256) ) if str(lowercase_ ).startswith("""mps""" ): lowercase_ : Union[str, Any] = torch.manual_seed(lowercase_ ) else: lowercase_ : Optional[int] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) lowercase_ : Dict = { """prompt""": """horse""", """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : Optional[Any] = """cpu""" lowercase_ : List[str] = self.get_dummy_components() lowercase_ : Optional[Any] = self.pipeline_class(**lowercase_ ) lowercase_ : List[Any] = pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) lowercase_ : Tuple = pipe(**self.get_dummy_inputs(lowercase_ ) ) lowercase_ : Optional[int] = output.images lowercase_ : List[str] = pipe( **self.get_dummy_inputs(lowercase_ ) , return_dict=lowercase_ , )[0] lowercase_ : Tuple = image[0, -3:, -3:, -1] lowercase_ : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : Dict = np.array( [0.61_47_49_43, 0.6_07_35_39, 0.43_30_85_44, 0.5_92_82_69, 0.47_49_35_95, 0.46_75_59_73, 0.4_61_38_38, 0.45_36_87_97, 0.50_11_92_33] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase): def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/kandinsky_img2img_frog.npy""" ) lowercase_ : Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) lowercase_ : int = """A red cartoon frog, 4k""" lowercase_ : List[Any] = KandinskyPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-1-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(lowercase_ ) lowercase_ : Optional[Any] = KandinskyImgaImgPipeline.from_pretrained( """kandinsky-community/kandinsky-2-1""" , torch_dtype=torch.floataa ) lowercase_ : Optional[Any] = pipeline.to(lowercase_ ) pipeline.set_progress_bar_config(disable=lowercase_ ) lowercase_ : str = torch.Generator(device="""cpu""" ).manual_seed(0 ) lowercase_ , lowercase_ : str = pipe_prior( lowercase_ , generator=lowercase_ , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() lowercase_ : Optional[int] = pipeline( lowercase_ , image=lowercase_ , image_embeds=lowercase_ , negative_image_embeds=lowercase_ , generator=lowercase_ , num_inference_steps=100 , height=768 , width=768 , strength=0.2 , output_type="""np""" , ) lowercase_ : Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowercase_ , lowercase_ )
239
'''simple docstring''' import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class __magic_name__ : def __init__( self : int , lowercase_ : Optional[int] , lowercase_ : Any=13 , lowercase_ : List[str]=7 , lowercase_ : List[Any]=True , lowercase_ : str=True , lowercase_ : Dict=True , lowercase_ : List[str]=True , lowercase_ : List[str]=99 , lowercase_ : Dict=32 , lowercase_ : List[Any]=5 , lowercase_ : List[str]=4 , lowercase_ : Dict=37 , lowercase_ : List[Any]="gelu" , lowercase_ : Dict=0.1 , lowercase_ : Any=0.1 , lowercase_ : int=512 , lowercase_ : Tuple=16 , lowercase_ : str=2 , lowercase_ : Optional[Any]=0.02 , lowercase_ : Any=3 , lowercase_ : Any=4 , lowercase_ : Dict=None , ): lowercase_ : Tuple = parent lowercase_ : Tuple = batch_size lowercase_ : Optional[int] = seq_length lowercase_ : Union[str, Any] = is_training lowercase_ : int = use_input_mask lowercase_ : Union[str, Any] = use_token_type_ids lowercase_ : Tuple = use_labels lowercase_ : Tuple = vocab_size lowercase_ : int = hidden_size lowercase_ : int = num_hidden_layers lowercase_ : Optional[int] = num_attention_heads lowercase_ : Union[str, Any] = intermediate_size lowercase_ : List[Any] = hidden_act lowercase_ : int = hidden_dropout_prob lowercase_ : Union[str, Any] = attention_probs_dropout_prob lowercase_ : List[Any] = max_position_embeddings lowercase_ : Union[str, Any] = type_vocab_size lowercase_ : List[Any] = type_sequence_label_size lowercase_ : Optional[int] = initializer_range lowercase_ : str = num_labels lowercase_ : int = num_choices lowercase_ : List[Any] = scope def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : str = None if self.use_input_mask: lowercase_ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : Optional[int] = None if self.use_token_type_ids: lowercase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase_ : str = None lowercase_ : Optional[int] = None lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : List[str] = ids_tensor([self.batch_size] , self.num_choices ) lowercase_ : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE_ ( self : int ): return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] , lowercase_ : Any , lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Tuple , lowercase_ : Optional[int] ): lowercase_ : Optional[Any] = NystromformerModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Optional[int] = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ ) lowercase_ : Optional[Any] = model(lowercase_ , token_type_ids=lowercase_ ) lowercase_ : Union[str, Any] = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE_ ( self : int , lowercase_ : Tuple , lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : List[Any] , lowercase_ : Optional[int] , lowercase_ : Any ): lowercase_ : List[Any] = NystromformerForMaskedLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Optional[int] = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE_ ( self : Tuple , lowercase_ : int , lowercase_ : List[Any] , lowercase_ : Any , lowercase_ : str , lowercase_ : Dict , lowercase_ : Any , lowercase_ : Tuple ): lowercase_ : Any = NystromformerForQuestionAnswering(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Union[str, Any] = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , start_positions=lowercase_ , end_positions=lowercase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE_ ( self : Tuple , lowercase_ : List[str] , lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : int ): lowercase_ : Any = self.num_labels lowercase_ : Union[str, Any] = NystromformerForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Any = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] , lowercase_ : List[str] , lowercase_ : int , lowercase_ : List[str] , lowercase_ : Dict , lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : List[str] ): lowercase_ : int = self.num_labels lowercase_ : int = NystromformerForTokenClassification(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Tuple = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] , lowercase_ : Tuple , lowercase_ : Dict , lowercase_ : int , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : Union[str, Any] ): lowercase_ : str = self.num_choices lowercase_ : Union[str, Any] = NystromformerForMultipleChoice(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : str = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowercase_ : Tuple = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowercase_ : Dict = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowercase_ : Union[str, Any] = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): lowercase_ : Union[str, Any] = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Tuple = config_and_inputs lowercase_ : Optional[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __magic_name__ ( _UpperCAmelCase, _UpperCAmelCase, unittest.TestCase): UpperCamelCase__ = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) UpperCamelCase__ = ( { '''feature-extraction''': NystromformerModel, '''fill-mask''': NystromformerForMaskedLM, '''question-answering''': NystromformerForQuestionAnswering, '''text-classification''': NystromformerForSequenceClassification, '''token-classification''': NystromformerForTokenClassification, '''zero-shot''': NystromformerForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : Any = NystromformerModelTester(self ) lowercase_ : Optional[Any] = ConfigTester(self , config_class=lowercase_ , hidden_size=37 ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE_ ( self : Dict ): lowercase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase_ : int = type self.model_tester.create_and_check_model(*lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase_ ) @slow def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : List[Any] = NystromformerModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @require_torch class __magic_name__ ( unittest.TestCase): @slow def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): lowercase_ : List[str] = NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) lowercase_ : List[str] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): lowercase_ : Tuple = model(lowercase_ )[0] lowercase_ : Tuple = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , lowercase_ ) lowercase_ : Dict = torch.tensor( [[[-0.45_32, -0.09_36, 0.51_37], [-0.26_76, 0.06_28, 0.61_86], [-0.36_29, -0.17_26, 0.47_16]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase_ , atol=1E-4 ) ) @slow def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : Optional[int] = """the [MASK] of Belgium is Brussels""" lowercase_ : Optional[Any] = AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) lowercase_ : List[Any] = NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) lowercase_ : str = tokenizer(lowercase_ , return_tensors="""pt""" ) with torch.no_grad(): lowercase_ : Tuple = model(encoding.input_ids ).logits lowercase_ : Optional[int] = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(lowercase_ ) , """capital""" )
239
1
"""simple docstring""" from __future__ import annotations A_ : int = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } class _lowerCAmelCase: """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase ): UpperCamelCase_: List[Any] = graph # mapping node to its parent in resulting breadth first tree UpperCamelCase_: dict[str, str | None] = {} UpperCamelCase_: Optional[Any] = source_vertex def _a ( self ): UpperCamelCase_: Optional[int] = {self.source_vertex} UpperCamelCase_: List[Any] = None UpperCamelCase_: Tuple = [self.source_vertex] # first in first out queue while queue: UpperCamelCase_: List[str] = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(_lowerCamelCase ) UpperCamelCase_: Tuple = vertex queue.append(_lowerCamelCase ) def _a ( self , _lowerCamelCase ): if target_vertex == self.source_vertex: return self.source_vertex UpperCamelCase_: Union[str, Any] = self.parent.get(_lowerCamelCase ) if target_vertex_parent is None: UpperCamelCase_: List[Any] = ( f'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(_lowerCamelCase ) return self.shortest_path(_lowerCamelCase ) + f'''->{target_vertex}''' if __name__ == "__main__": A_ : Optional[int] = Graph(graph, 'G') g.breath_first_search() print(g.shortest_path('D')) print(g.shortest_path('G')) print(g.shortest_path('Foo'))
360
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class _lowerCAmelCase( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" a : Optional[int] =RoFormerTokenizer a : int =RoFormerTokenizerFast a : int =True a : Optional[int] =True def _a ( self ): super().setUp() def _a ( self , **_lowerCamelCase ): return self.tokenizer_class.from_pretrained('junnyu/roformer_chinese_base' , **_lowerCamelCase ) def _a ( self , **_lowerCamelCase ): return self.rust_tokenizer_class.from_pretrained('junnyu/roformer_chinese_base' , **_lowerCamelCase ) def _a ( self ): UpperCamelCase_: Optional[int] = '永和服装饰品有限公司,今天天气非常好' UpperCamelCase_: Any = '永和 服装 饰品 有限公司 , 今 天 天 气 非常 好' return input_text, output_text def _a ( self ): UpperCamelCase_: int = self.get_tokenizer() UpperCamelCase_ ,UpperCamelCase_: int = self.get_chinese_input_output_texts() UpperCamelCase_: Tuple = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , output_text.split() ) UpperCamelCase_: Dict = tokens + [tokenizer.unk_token] UpperCamelCase_: Dict = [2_2_9_4_3, 2_1_3_3_2, 3_4_4_3_1, 4_5_9_0_4, 1_1_7, 3_0_6, 1_2_3_1, 1_2_3_1, 2_6_5_3, 3_3_9_9_4, 1_2_6_6, 1_0_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , _lowerCamelCase ) def _a ( self ): UpperCamelCase_: Optional[Any] = self.get_rust_tokenizer() UpperCamelCase_ ,UpperCamelCase_: Tuple = self.get_chinese_input_output_texts() UpperCamelCase_: Optional[Any] = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , output_text.split() ) UpperCamelCase_: str = tokens + [tokenizer.unk_token] UpperCamelCase_: Optional[Any] = [2_2_9_4_3, 2_1_3_3_2, 3_4_4_3_1, 4_5_9_0_4, 1_1_7, 3_0_6, 1_2_3_1, 1_2_3_1, 2_6_5_3, 3_3_9_9_4, 1_2_6_6, 1_0_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , _lowerCamelCase ) def _a ( self ): pass def _a ( self ): pass def _a ( self ): pass
292
0
'''simple docstring''' # flake8: noqa # Lint as: python3 a_ : Optional[int] = [ """VerificationMode""", """Version""", """disable_progress_bar""", """enable_progress_bar""", """is_progress_bar_enabled""", """experimental""", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
75
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets A__ = datasets.logging.get_logger(__name__) A__ = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' A__ = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' A__ = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=True , __lowerCAmelCase=False , __lowerCAmelCase="dummy_doc" ) -> int: """simple docstring""" snake_case__ : Dict = {doc: key_lines} snake_case__ : Any = {doc: sys_lines} snake_case__ : Dict = {} snake_case__ : List[str] = 0 snake_case__ : Optional[Any] = 0 snake_case__ : Optional[Any] = 0 snake_case__ : Dict = 0 snake_case__ : List[Any] = 0 snake_case__ : List[Any] = 0 snake_case__ , snake_case__ : Tuple = reader.get_doc_mentions(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: snake_case__ : str = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ , snake_case__ : int = reader.get_doc_mentions(__lowerCAmelCase , sys_doc_lines[doc] , __lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: snake_case__ : Union[str, Any] = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) if remove_nested: snake_case__ , snake_case__ : Dict = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters snake_case__ , snake_case__ : Optional[int] = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters snake_case__ : Any = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : Optional[int] = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : List[Any] = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( '''Number of resulting singleton clusters in the key ''' f"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( f"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ '''files, respectively''' ) return doc_coref_infos def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: """simple docstring""" snake_case__ : Optional[Any] = get_coref_infos(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ : str = {} snake_case__ : Optional[int] = 0 snake_case__ : List[Any] = 0 for name, metric in metrics: snake_case__ , snake_case__ , snake_case__ : Any = evaluator.evaluate_documents(__lowerCAmelCase , __lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"""{name}/recall""": recall, f"""{name}/precision""": precision, f"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , f"""Recall: {recall * 100:.2f}""" , f""" Precision: {precision * 100:.2f}""" , f""" F1: {fa * 100:.2f}""" , ) if conll_subparts_num == 3: snake_case__ : int = (conll / 3) * 100 logger.info(f"""CoNLL score: {conll:.2f}""" ) output_scores.update({'''conll_score''': conll} ) return output_scores def _lowerCAmelCase ( __lowerCAmelCase ) -> List[str]: """simple docstring""" snake_case__ : str = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: snake_case__ : List[Any] = line.split()[5] if not parse_col == "-": snake_case__ : Optional[Any] = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): def __lowerCamelCase ( self :Dict ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) ,codebase_urls=['''https://github.com/ns-moosavi/coval'''] ,reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] ,) def __lowerCamelCase ( self :Any ,__lowercase :List[Any] ,__lowercase :int ,__lowercase :str=True ,__lowercase :Optional[int]=False ,__lowercase :Optional[Any]=False ,__lowercase :Tuple=False ): snake_case__ : Optional[Any] = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: snake_case__ : Optional[int] = util.check_gold_parse_annotation(__lowercase ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" snake_case__ : Any = evaluate( key_lines=__lowercase ,sys_lines=__lowercase ,metrics=__lowercase ,NP_only=__lowercase ,remove_nested=__lowercase ,keep_singletons=__lowercase ,min_span=__lowercase ,) return score
230
0
"""simple docstring""" import inspect import unittest from transformers import ViTMSNConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMSNForImageClassification, ViTMSNModel from transformers.models.vit_msn.modeling_vit_msn import VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=30 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=None , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope # in ViT MSN, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def a__ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, pixel_values, labels def a__ ( self ): return ViTMSNConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTMSNModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() __a = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.type_sequence_label_size __a = ViTMSNForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() __a = model(lowercase_ , labels=lowercase_ ) print("Pixel and labels shape: {pixel_values.shape}, {labels.shape}" ) print("Labels: {labels}" ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a = 1 __a = ViTMSNForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ): __a = self.prepare_config_and_inputs() __a , __a , __a = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, unittest.TestCase ): _snake_case : int = (ViTMSNModel, ViTMSNForImageClassification) if is_torch_available() else () _snake_case : Optional[int] = ( {"""feature-extraction""": ViTMSNModel, """image-classification""": ViTMSNForImageClassification} if is_torch_available() else {} ) _snake_case : Optional[int] = False _snake_case : Optional[Any] = False _snake_case : List[Any] = False _snake_case : List[Any] = False def a__ ( self ): __a = ViTMSNModelTester(self ) __a = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViTMSN does not use inputs_embeds" ) def a__ ( self ): pass def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowercase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_ , nn.Linear ) ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowercase_ ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) @slow def a__ ( self ): for model_name in VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ViTMSNModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def _lowerCamelCase( ): __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def a__ ( self ): return ViTImageProcessor.from_pretrained("facebook/vit-msn-small" ) if is_vision_available() else None @slow def a__ ( self ): torch.manual_seed(2 ) __a = ViTMSNForImageClassification.from_pretrained("facebook/vit-msn-small" ).to(lowercase_ ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowercase_ , return_tensors="pt" ).to(lowercase_ ) # forward pass with torch.no_grad(): __a = model(**lowercase_ ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowercase_ ) __a = torch.tensor([-0.0803, -0.4454, -0.2375] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1E-4 ) )
369
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: SCREAMING_SNAKE_CASE__:List[Any] = None SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Optional[int] = { """vocab_file""": { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model""", }, """tokenizer_file""": { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__:List[str] = { """camembert-base""": 512, } SCREAMING_SNAKE_CASE__:str = """▁""" class snake_case__ ( snake_case_ ): _snake_case : List[Any] = VOCAB_FILES_NAMES _snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : Any = ["""input_ids""", """attention_mask"""] _snake_case : str = CamembertTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="</s>" , lowerCamelCase="<s>" , lowerCamelCase="<unk>" , lowerCamelCase="<pad>" , lowerCamelCase="<mask>" , lowerCamelCase=["<s>NOTUSED", "</s>NOTUSED"] , **lowerCamelCase , ): # Mask token behave like a normal word, i.e. include the space before it __a = AddedToken(lowerCamelCase , lstrip=lowerCamelCase , rstrip=lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ) else mask_token super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , sep_token=lowerCamelCase , cls_token=lowerCamelCase , unk_token=lowerCamelCase , pad_token=lowerCamelCase , mask_token=lowerCamelCase , additional_special_tokens=lowerCamelCase , **lowerCamelCase , ) __a = vocab_file __a = False if not self.vocab_file else True def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __a = [self.cls_token_id] __a = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCamelCase ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __a = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ): copyfile(self.vocab_file , lowerCamelCase ) return (out_vocab_file,)
268
0
'''simple docstring''' def a ( lowerCamelCase__ = "The quick brown fox jumps over the lazy dog" , ): '''simple docstring''' A_ : Any = set() # Replace all the whitespace in our sentence A_ : Optional[Any] = input_str.replace(""" """ , """""" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(lowerCamelCase__ ) == 26 def a ( lowerCamelCase__ = "The quick brown fox jumps over the lazy dog" , ): '''simple docstring''' A_ : List[Any] = [False] * 26 for char in input_str: if char.islower(): A_ : Optional[int] = True elif char.isupper(): A_ : Any = True return all(lowerCamelCase__ ) def a ( lowerCamelCase__ = "The quick brown fox jumps over the lazy dog" , ): '''simple docstring''' return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def a ( ): '''simple docstring''' from timeit import timeit A_ : Union[str, Any] = """from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest""" print(timeit("""is_pangram()""" , setup=lowerCamelCase__ ) ) print(timeit("""is_pangram_faster()""" , setup=lowerCamelCase__ ) ) print(timeit("""is_pangram_fastest()""" , setup=lowerCamelCase__ ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
206
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() lowerCamelCase : Any = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ) -> Any: snake_case : Optional[int] = WavaVecaForSequenceClassification.from_pretrained(lowercase ,config=lowercase ) snake_case : List[str] = downstream_dict["""projector.weight"""] snake_case : Dict = downstream_dict["""projector.bias"""] snake_case : Dict = downstream_dict["""model.post_net.linear.weight"""] snake_case : List[Any] = downstream_dict["""model.post_net.linear.bias"""] return model def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ) -> List[str]: snake_case : str = WavaVecaForAudioFrameClassification.from_pretrained(lowercase ,config=lowercase ) snake_case : List[Any] = downstream_dict["""model.linear.weight"""] snake_case : str = downstream_dict["""model.linear.bias"""] return model def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ) -> str: snake_case : Any = WavaVecaForXVector.from_pretrained(lowercase ,config=lowercase ) snake_case : str = downstream_dict["""connector.weight"""] snake_case : Optional[Any] = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): snake_case : List[Any] = downstream_dict[ f"""model.framelevel_feature_extractor.module.{i}.kernel.weight""" ] snake_case : Optional[int] = downstream_dict[f"""model.framelevel_feature_extractor.module.{i}.kernel.bias"""] snake_case : List[str] = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] snake_case : Union[str, Any] = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] snake_case : Any = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] snake_case : int = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] snake_case : Any = downstream_dict["""objective.W"""] return model @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ) -> Union[str, Any]: snake_case : Tuple = torch.load(lowercase ,map_location="""cpu""" ) snake_case : Any = checkpoint["""Downstream"""] snake_case : List[str] = WavaVecaConfig.from_pretrained(lowercase ) snake_case : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained( lowercase ,return_attention_mask=lowercase ,do_normalize=lowercase ) snake_case : str = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): snake_case : int = convert_classification(lowercase ,lowercase ,lowercase ) elif arch.endswith("""ForAudioFrameClassification""" ): snake_case : Dict = convert_diarization(lowercase ,lowercase ,lowercase ) elif arch.endswith("""ForXVector""" ): snake_case : Optional[Any] = convert_xvector(lowercase ,lowercase ,lowercase ) else: raise NotImplementedError(f"""S3PRL weights conversion is not supported for {arch}""" ) if hf_config.use_weighted_layer_sum: snake_case : List[str] = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(lowercase ) hf_model.save_pretrained(lowercase ) if __name__ == "__main__": lowerCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') lowerCamelCase : int = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
124
0
from __future__ import annotations from collections import deque class __lowerCAmelCase : def __init__( self: Union[str, Any] , _lowerCAmelCase: list[str] ): lowercase :list[dict] = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__a ) self.set_fail_transitions() def SCREAMING_SNAKE_CASE ( self: str , _lowerCAmelCase: int , _lowerCAmelCase: str ): for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: str ): lowercase :List[str] = 0 for character in keyword: lowercase :Any = self.find_next_state(__a , __a ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) lowercase :Optional[int] = len(self.adlist ) - 1 else: lowercase :List[str] = next_state self.adlist[current_state]["output"].append(__a ) def SCREAMING_SNAKE_CASE ( self: List[str] ): lowercase :deque = deque() for node in self.adlist[0]["next_states"]: q.append(__a ) lowercase :Optional[int] = 0 while q: lowercase :Optional[int] = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__a ) lowercase :Optional[int] = self.adlist[r]['fail_state'] while ( self.find_next_state(__a , self.adlist[child]["value"] ) is None and state != 0 ): lowercase :Optional[int] = self.adlist[state]['fail_state'] lowercase :int = self.find_next_state( __a , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: lowercase :Optional[int] = 0 lowercase :Dict = ( self.adlist[child]['output'] + self.adlist[self.adlist[child]['fail_state']]['output'] ) def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: str ): lowercase :dict = {} # returns a dict with keywords and list of its occurrences lowercase :Tuple = 0 for i in range(len(__a ) ): while ( self.find_next_state(__a , string[i] ) is None and current_state != 0 ): lowercase :Dict = self.adlist[current_state]['fail_state'] lowercase :Union[str, Any] = self.find_next_state(__a , string[i] ) if next_state is None: lowercase :Union[str, Any] = 0 else: lowercase :Optional[Any] = next_state for key in self.adlist[current_state]["output"]: if key not in result: lowercase :Tuple = [] result[key].append(i - len(__a ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
350
import pytest _UpperCAmelCase : List[Any] = "__dummy_dataset1__" _UpperCAmelCase : Union[str, Any] = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def UpperCAmelCase__ ( ): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def UpperCAmelCase__ ( ): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowercase :Tuple = dataset_loading_script_name lowercase :Dict = tmp_path / "datasets" / script_name script_dir.mkdir(parents=lowerCamelCase ) lowercase :int = script_dir / F"{script_name}.py" with open(lowerCamelCase, "w" ) as f: f.write(lowerCamelCase ) return str(lowerCamelCase )
158
0
"""simple docstring""" from __future__ import annotations UpperCAmelCase__ : Tuple = [] def lowercase_ ( _snake_case ,_snake_case ,_snake_case ): for i in range(len(_snake_case ) ): if board[row][i] == 1: return False for i in range(len(_snake_case ) ): if board[i][column] == 1: return False for i, j in zip(range(_snake_case ,-1 ,-1 ) ,range(_snake_case ,-1 ,-1 ) ): if board[i][j] == 1: return False for i, j in zip(range(_snake_case ,-1 ,-1 ) ,range(_snake_case ,len(_snake_case ) ) ): if board[i][j] == 1: return False return True def lowercase_ ( _snake_case ,_snake_case ): if row >= len(_snake_case ): solution.append(_snake_case ) printboard(_snake_case ) print() return True for i in range(len(_snake_case ) ): if is_safe(_snake_case ,_snake_case ,_snake_case ): SCREAMING_SNAKE_CASE__ : List[str] = 1 solve(_snake_case ,row + 1 ) SCREAMING_SNAKE_CASE__ : int = 0 return False def lowercase_ ( _snake_case ): for i in range(len(_snake_case ) ): for j in range(len(_snake_case ) ): if board[i][j] == 1: print("""Q""" ,end=""" """ ) else: print(""".""" ,end=""" """ ) print() # n=int(input("The no. of queens")) UpperCAmelCase__ : Optional[Any] = 8 UpperCAmelCase__ : Tuple = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('The total no. of solutions are :', len(solution))
25
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging a__ : List[Any] = logging.get_logger(__name__) # TODO: upload to AWS a__ : List[str] = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json" ), } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE): UpperCAmelCase__ : Union[str, Any] = 'retribert' def __init__( self :int , _A :str=30_522 , _A :Optional[int]=768 , _A :List[Any]=8 , _A :Tuple=12 , _A :Optional[int]=3_072 , _A :Union[str, Any]="gelu" , _A :List[str]=0.1 , _A :Tuple=0.1 , _A :List[Any]=512 , _A :Dict=2 , _A :Optional[int]=0.02 , _A :List[str]=1E-12 , _A :Optional[int]=True , _A :int=128 , _A :Tuple=0 , **_A :str , ) -> str: '''simple docstring''' super().__init__(pad_token_id=_A , **_A ) __A = vocab_size __A = hidden_size __A = num_hidden_layers __A = num_attention_heads __A = hidden_act __A = intermediate_size __A = hidden_dropout_prob __A = attention_probs_dropout_prob __A = max_position_embeddings __A = type_vocab_size __A = initializer_range __A = layer_norm_eps __A = share_encoders __A = projection_dim
161
0
"""simple docstring""" from __future__ import annotations __A =[-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __A =[-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = [] lowerCamelCase_ = len(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): lowerCamelCase_ = -1 for j in range(i + 1 , lowerCamelCase__ ): if arr[i] < arr[j]: lowerCamelCase_ = arr[j] break result.append(lowerCamelCase__ ) return result def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = [] for i, outer in enumerate(lowerCamelCase__ ): lowerCamelCase_ = -1 for inner in arr[i + 1 :]: if outer < inner: lowerCamelCase_ = inner break result.append(lowerCamelCase__ ) return result def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = len(lowerCamelCase__ ) lowerCamelCase_ = [] lowerCamelCase_ = [-1] * arr_size for index in reversed(range(lowerCamelCase__ ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: lowerCamelCase_ = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __A =( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
359
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class _SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self , lowercase = 16 , lowercase = 88 , lowercase = None , lowercase = 1 , lowercase = 0.0 , lowercase = 32 , lowercase = None , lowercase = False , lowercase = None , lowercase = None , lowercase = "geglu" , lowercase = None , ) -> Any: super().__init__() lowerCamelCase_ = nn.ModuleList( [ TransformeraDModel( num_attention_heads=lowercase , attention_head_dim=lowercase , in_channels=lowercase , num_layers=lowercase , dropout=lowercase , norm_num_groups=lowercase , cross_attention_dim=lowercase , attention_bias=lowercase , sample_size=lowercase , num_vector_embeds=lowercase , activation_fn=lowercase , num_embeds_ada_norm=lowercase , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference lowerCamelCase_ = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` lowerCamelCase_ = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` lowerCamelCase_ = [1, 0] def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase=None , lowercase=None , lowercase=None , lowercase = True , ) -> int: lowerCamelCase_ = hidden_states lowerCamelCase_ = [] lowerCamelCase_ = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens lowerCamelCase_ = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] lowerCamelCase_ = self.transformer_index_for_condition[i] lowerCamelCase_ = self.transformers[transformer_index]( lowercase , encoder_hidden_states=lowercase , timestep=lowercase , cross_attention_kwargs=lowercase , return_dict=lowercase , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] lowerCamelCase_ = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) lowerCamelCase_ = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=lowercase )
47
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCamelCase = { '''configuration_x_clip''': [ '''XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XCLIPConfig''', '''XCLIPTextConfig''', '''XCLIPVisionConfig''', ], '''processing_x_clip''': ['''XCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XCLIPModel''', '''XCLIPPreTrainedModel''', '''XCLIPTextModel''', '''XCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
254
'''simple docstring''' import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline _UpperCamelCase = version.parse(version.parse(torch.__version__).base_version) < version.parse('''1.11''') def lowercase_ ( lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : tuple , lowerCAmelCase__ : Path , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[int]=False , ): """simple docstring""" output_path.parent.mkdir(parents=lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , use_external_data_format=lowerCAmelCase__ , enable_onnx_checker=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) else: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) @torch.no_grad() def lowercase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : str , lowerCAmelCase__ : int , lowerCAmelCase__ : bool = False ): """simple docstring""" __UpperCAmelCase : Tuple = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): __UpperCAmelCase : Optional[int] = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: __UpperCAmelCase : Dict = """cpu""" __UpperCAmelCase : Optional[Any] = StableDiffusionPipeline.from_pretrained(lowerCAmelCase__ , torch_dtype=lowerCAmelCase__ ).to(lowerCAmelCase__ ) __UpperCAmelCase : List[str] = Path(lowerCAmelCase__ ) # TEXT ENCODER __UpperCAmelCase : Any = pipeline.text_encoder.config.max_position_embeddings __UpperCAmelCase : str = pipeline.text_encoder.config.hidden_size __UpperCAmelCase : Optional[Any] = pipeline.tokenizer( """A sample prompt""" , padding="""max_length""" , max_length=pipeline.tokenizer.model_max_length , truncation=lowerCAmelCase__ , return_tensors="""pt""" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=lowerCAmelCase__ , dtype=torch.intaa )) , output_path=output_path / """text_encoder""" / """model.onnx""" , ordered_input_names=["""input_ids"""] , output_names=["""last_hidden_state""", """pooler_output"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase__ , ) del pipeline.text_encoder # UNET __UpperCAmelCase : Optional[int] = pipeline.unet.config.in_channels __UpperCAmelCase : Tuple = pipeline.unet.config.sample_size __UpperCAmelCase : Dict = output_path / """unet""" / """model.onnx""" onnx_export( pipeline.unet , model_args=( torch.randn(2 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), torch.randn(2 ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), torch.randn(2 , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=lowerCAmelCase__ , ordered_input_names=["""sample""", """timestep""", """encoder_hidden_states""", """return_dict"""] , output_names=["""out_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """timestep""": {0: """batch"""}, """encoder_hidden_states""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase__ , use_external_data_format=lowerCAmelCase__ , ) __UpperCAmelCase : Any = str(unet_path.absolute().as_posix() ) __UpperCAmelCase : int = os.path.dirname(lowerCAmelCase__ ) __UpperCAmelCase : Tuple = onnx.load(lowerCAmelCase__ ) # clean up existing tensor files shutil.rmtree(lowerCAmelCase__ ) os.mkdir(lowerCAmelCase__ ) # collate external tensor files into one onnx.save_model( lowerCAmelCase__ , lowerCAmelCase__ , save_as_external_data=lowerCAmelCase__ , all_tensors_to_one_file=lowerCAmelCase__ , location="""weights.pb""" , convert_attribute=lowerCAmelCase__ , ) del pipeline.unet # VAE ENCODER __UpperCAmelCase : Union[str, Any] = pipeline.vae __UpperCAmelCase : str = vae_encoder.config.in_channels __UpperCAmelCase : Any = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder __UpperCAmelCase : str = lambda lowerCAmelCase__ , lowerCAmelCase__ : vae_encoder.encode(lowerCAmelCase__ , lowerCAmelCase__ )[0].sample() onnx_export( lowerCAmelCase__ , model_args=( torch.randn(1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=output_path / """vae_encoder""" / """model.onnx""" , ordered_input_names=["""sample""", """return_dict"""] , output_names=["""latent_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase__ , ) # VAE DECODER __UpperCAmelCase : Optional[Any] = pipeline.vae __UpperCAmelCase : Optional[int] = vae_decoder.config.latent_channels __UpperCAmelCase : Dict = vae_decoder.config.out_channels # forward only through the decoder part __UpperCAmelCase : List[Any] = vae_encoder.decode onnx_export( lowerCAmelCase__ , model_args=( torch.randn(1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase__ , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: __UpperCAmelCase : Tuple = pipeline.safety_checker __UpperCAmelCase : Union[str, Any] = safety_checker.config.vision_config.num_channels __UpperCAmelCase : Any = safety_checker.config.vision_config.image_size __UpperCAmelCase : Optional[int] = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), torch.randn(1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), ) , output_path=output_path / """safety_checker""" / """model.onnx""" , ordered_input_names=["""clip_input""", """images"""] , output_names=["""out_images""", """has_nsfw_concepts"""] , dynamic_axes={ """clip_input""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """images""": {0: """batch""", 1: """height""", 2: """width""", 3: """channels"""}, } , opset=lowerCAmelCase__ , ) del pipeline.safety_checker __UpperCAmelCase : Optional[Any] = OnnxRuntimeModel.from_pretrained(output_path / """safety_checker""" ) __UpperCAmelCase : Any = pipeline.feature_extractor else: __UpperCAmelCase : List[str] = None __UpperCAmelCase : Any = None __UpperCAmelCase : Tuple = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_encoder""" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_decoder""" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / """text_encoder""" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / """unet""" ) , scheduler=pipeline.scheduler , safety_checker=lowerCAmelCase__ , feature_extractor=lowerCAmelCase__ , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(lowerCAmelCase__ ) print("""ONNX pipeline saved to""" , lowerCAmelCase__ ) del pipeline del onnx_pipeline __UpperCAmelCase : Tuple = OnnxStableDiffusionPipeline.from_pretrained(lowerCAmelCase__ , provider="""CPUExecutionProvider""" ) print("""ONNX pipeline is loadable""" ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--model_path''', type=str, required=True, help='''Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).''', ) parser.add_argument('''--output_path''', type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--opset''', default=14, type=int, help='''The version of the ONNX operator set to use.''', ) parser.add_argument('''--fp16''', action='''store_true''', default=False, help='''Export the models in `float16` mode''') _UpperCamelCase = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
254
1
"""simple docstring""" def a__ ( __lowercase=2_8123 ) -> List[Any]: _A = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i _A = set() _A = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(__lowercase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
355
"""simple docstring""" def a__ ( __lowercase ) -> int: assert ( isinstance(__lowercase , __lowercase ) and number_of_steps > 0 ), f"""number_of_steps needs to be positive integer, your input {number_of_steps}""" if number_of_steps == 1: return 1 _A , _A = 1, 1 for _ in range(number_of_steps - 1 ): _A , _A = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
163
0
'''simple docstring''' import math import sys def UpperCamelCase_( snake_case : int ): '''simple docstring''' if number != int(snake_case ): raise ValueError("the value of input must be a natural number" ) if number < 0: raise ValueError("the value of input must not be a negative number" ) if number == 0: return 1 snake_case_ = [-1] * (number + 1) snake_case_ = 0 for i in range(1 , number + 1 ): snake_case_ = sys.maxsize snake_case_ = int(math.sqrt(snake_case ) ) for j in range(1 , root + 1 ): snake_case_ = 1 + answers[i - (j**2)] snake_case_ = min(snake_case , snake_case ) snake_case_ = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
85
'''simple docstring''' from statistics import mean, stdev def UpperCamelCase_( snake_case : list , snake_case : int = 3 ): '''simple docstring''' snake_case_ = min(snake_case ) snake_case_ = max(snake_case ) # normalize data return [round((x - x_min) / (x_max - x_min) , snake_case ) for x in data] def UpperCamelCase_( snake_case : list , snake_case : int = 3 ): '''simple docstring''' snake_case_ = mean(snake_case ) snake_case_ = stdev(snake_case ) # standardize data return [round((x - mu) / (sigma) , snake_case ) for x in data]
85
1
def A ( a_ ,a_ ) -> str: __UpperCamelCase : Union[str, Any] ='' for i in table: res += inp[i - 1] return res def A ( a_ ) -> str: return data[1:] + data[0] def A ( a_ ,a_ ) -> Dict: __UpperCamelCase : int ='' for i in range(len(a_ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def A ( a_ ,a_ ) -> Tuple: __UpperCamelCase : List[str] =int('0b' + data[0] + data[-1] ,2 ) __UpperCamelCase : List[str] =int('0b' + data[1:3] ,2 ) return bin(s[row][col] )[2:] def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Optional[int]: __UpperCamelCase : List[Any] =message[:4] __UpperCamelCase : Tuple =message[4:] __UpperCamelCase : Any =apply_table(a_ ,a_ ) __UpperCamelCase : List[Any] =xor(a_ ,a_ ) __UpperCamelCase : Union[str, Any] =apply_sbox(a_ ,temp[:4] ) # noqa: E741 __UpperCamelCase : str =apply_sbox(a_ ,temp[4:] ) __UpperCamelCase : Any ='0' * (2 - len(a_ )) + l # noqa: E741 __UpperCamelCase : Union[str, Any] ='0' * (2 - len(a_ )) + r __UpperCamelCase : Union[str, Any] =apply_table(l + r ,a_ ) __UpperCamelCase : Optional[Any] =xor(a_ ,a_ ) return temp + right if __name__ == "__main__": A_ :int = input('''Enter 10 bit key: ''') A_ :Tuple = input('''Enter 8 bit message: ''') A_ :int = [6, 3, 7, 4, 8, 5, 10, 9] A_ :int = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] A_ :Union[str, Any] = [2, 4, 3, 1] A_ :Dict = [2, 6, 3, 1, 4, 8, 5, 7] A_ :Any = [4, 1, 3, 5, 7, 2, 8, 6] A_ :int = [4, 1, 2, 3, 2, 3, 4, 1] A_ :int = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] A_ :Dict = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation A_ :int = apply_table(key, paa_table) A_ :int = temp[:5] A_ :List[Any] = temp[5:] A_ :Optional[Any] = left_shift(left) A_ :List[str] = left_shift(right) A_ :List[str] = apply_table(left + right, pa_table) A_ :Dict = left_shift(left) A_ :Tuple = left_shift(right) A_ :Dict = left_shift(left) A_ :Union[str, Any] = left_shift(right) A_ :Optional[Any] = apply_table(left + right, pa_table) # encryption A_ :Optional[Any] = apply_table(message, IP) A_ :Tuple = function(expansion, sa, sa, keya, temp) A_ :List[str] = temp[4:] + temp[:4] A_ :Union[str, Any] = function(expansion, sa, sa, keya, temp) A_ :str = apply_table(temp, IP_inv) print('''Cipher text is:''', CT) # decryption A_ :List[Any] = apply_table(CT, IP) A_ :int = function(expansion, sa, sa, keya, temp) A_ :Optional[int] = temp[4:] + temp[:4] A_ :Optional[int] = function(expansion, sa, sa, keya, temp) A_ :Any = apply_table(temp, IP_inv) print('''Plain text after decypting is:''', PT)
245
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" return f'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase__ ) for s in shape] )}.npy' def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() def __lowercase ( self , lowerCamelCase__=0 , lowerCamelCase__=(4, 4, 64, 64) , lowerCamelCase__=False ): """simple docstring""" __UpperCamelCase : str =jnp.bfloataa if fpaa else jnp.floataa __UpperCamelCase : Optional[Any] =jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase__ , lowerCamelCase__ ) ) , dtype=lowerCamelCase__ ) return image def __lowercase ( self , lowerCamelCase__=False , lowerCamelCase__="CompVis/stable-diffusion-v1-4" ): """simple docstring""" __UpperCamelCase : List[Any] =jnp.bfloataa if fpaa else jnp.floataa __UpperCamelCase : Optional[int] ='bf16' if fpaa else None __UpperCamelCase , __UpperCamelCase : Any =FlaxUNetaDConditionModel.from_pretrained( lowerCamelCase__ , subfolder='unet' , dtype=lowerCamelCase__ , revision=lowerCamelCase__ ) return model, params def __lowercase ( self , lowerCamelCase__=0 , lowerCamelCase__=(4, 77, 768) , lowerCamelCase__=False ): """simple docstring""" __UpperCamelCase : str =jnp.bfloataa if fpaa else jnp.floataa __UpperCamelCase : Optional[int] =jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase__ , lowerCamelCase__ ) ) , dtype=lowerCamelCase__ ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_323, -0.1_304, 0.0_813, -0.3_093, -0.0_919, -0.1_571, -0.1_125, -0.5_806]], [17, 0.55, [-0.0_831, -0.2_443, 0.0_901, -0.0_919, 0.3_396, 0.0_103, -0.3_743, 0.0_701]], [8, 0.89, [-0.4_863, 0.0_859, 0.0_875, -0.1_658, 0.9_199, -0.0_114, 0.4_839, 0.4_639]], [3, 1000, [-0.5_649, 0.2_402, -0.5_518, 0.1_248, 1.1_328, -0.2_443, -0.0_325, -1.0_078]], # fmt: on ] ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Dict =self.get_unet_model(model_id='CompVis/stable-diffusion-v1-4' , fpaa=lowerCamelCase__ ) __UpperCamelCase : Dict =self.get_latents(lowerCamelCase__ , fpaa=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =self.get_encoder_hidden_states(lowerCamelCase__ , fpaa=lowerCamelCase__ ) __UpperCamelCase : List[str] =model.apply( {'params': params} , lowerCamelCase__ , jnp.array(lowerCamelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=lowerCamelCase__ , ).sample assert sample.shape == latents.shape __UpperCamelCase : List[str] =jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) __UpperCamelCase : int =jnp.array(lowerCamelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_514, 0.0_807, 0.1_624, 0.1_016, -0.1_896, 0.0_263, 0.0_677, 0.2_310]], [17, 0.55, [0.1_164, -0.0_216, 0.0_170, 0.1_589, -0.3_120, 0.1_005, -0.0_581, -0.1_458]], [8, 0.89, [-0.1_758, -0.0_169, 0.1_004, -0.1_411, 0.1_312, 0.1_103, -0.1_996, 0.2_139]], [3, 1000, [0.1_214, 0.0_352, -0.0_731, -0.1_562, -0.0_994, -0.0_906, -0.2_340, -0.0_539]], # fmt: on ] ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Dict =self.get_unet_model(model_id='stabilityai/stable-diffusion-2' , fpaa=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =self.get_latents(lowerCamelCase__ , shape=(4, 4, 96, 96) , fpaa=lowerCamelCase__ ) __UpperCamelCase : int =self.get_encoder_hidden_states(lowerCamelCase__ , shape=(4, 77, 1024) , fpaa=lowerCamelCase__ ) __UpperCamelCase : str =model.apply( {'params': params} , lowerCamelCase__ , jnp.array(lowerCamelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=lowerCamelCase__ , ).sample assert sample.shape == latents.shape __UpperCamelCase : int =jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) __UpperCamelCase : Optional[Any] =jnp.array(lowerCamelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-2 )
245
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : str =logging.get_logger(__name__) lowerCAmelCase__ : Union[str, Any] ={} class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ : Any = '''llama''' UpperCamelCase__ : Tuple = ['''past_key_values'''] def __init__( self , _A=32_000 , _A=4_096 , _A=11_008 , _A=32 , _A=32 , _A=None , _A="silu" , _A=2_048 , _A=0.0_2 , _A=1e-6 , _A=True , _A=0 , _A=1 , _A=2 , _A=1 , _A=False , _A=None , **_A , ): '''simple docstring''' __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads # for backward compatibility if num_key_value_heads is None: __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = num_key_value_heads __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = rms_norm_eps __SCREAMING_SNAKE_CASE = pretraining_tp __SCREAMING_SNAKE_CASE = use_cache __SCREAMING_SNAKE_CASE = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , tie_word_embeddings=_A , **_A , ) def _A ( self ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _A ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' f"""got {self.rope_scaling}""" ) __SCREAMING_SNAKE_CASE = self.rope_scaling.get('type' , _A ) __SCREAMING_SNAKE_CASE = self.rope_scaling.get('factor' , _A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(_A , _A ) or rope_scaling_factor <= 1.0: raise ValueError(f"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
257
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCAmelCase__ : List[Any] =input('''Enter image url: ''').strip() print(F'''Downloading image from {url} ...''') lowerCAmelCase__ : int =BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image lowerCAmelCase__ : Union[str, Any] =soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] lowerCAmelCase__ : int =requests.get(image_url).content lowerCAmelCase__ : Optional[int] =F'''{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg''' with open(file_name, '''wb''') as fp: fp.write(image_data) print(F'''Done. Image saved to disk as {file_name}.''')
257
1
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = inspect.getfile(accelerate.test_utils ) UpperCAmelCase = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) UpperCAmelCase = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) UpperCAmelCase = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def UpperCamelCase_ ( self ) -> int: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) UpperCAmelCase = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(snake_case__ , env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) UpperCAmelCase = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(snake_case__ , env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(snake_case__ , env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase_ ( self ) -> Any: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) UpperCAmelCase = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(snake_case__ , env=os.environ.copy() ) if __name__ == "__main__": lowerCAmelCase_ : Optional[int] = Accelerator() lowerCAmelCase_ : List[str] = (accelerator.state.process_index + 2, 1_0) lowerCAmelCase_ : str = torch.randint(0, 1_0, shape).to(accelerator.device) lowerCAmelCase_ : str = '''''' lowerCAmelCase_ : int = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." lowerCAmelCase_ : List[str] = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." lowerCAmelCase_ : Optional[Any] = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
248
"""simple docstring""" import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline lowerCAmelCase_ : Tuple = version.parse(version.parse(torch.__version__).base_version) < version.parse('''1.11''') def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False , ): '''simple docstring''' output_path.parent.mkdir(parents=lowerCAmelCase , exist_ok=lowerCAmelCase ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowerCAmelCase , lowerCAmelCase , f=output_path.as_posix() , input_names=lowerCAmelCase , output_names=lowerCAmelCase , dynamic_axes=lowerCAmelCase , do_constant_folding=lowerCAmelCase , use_external_data_format=lowerCAmelCase , enable_onnx_checker=lowerCAmelCase , opset_version=lowerCAmelCase , ) else: export( lowerCAmelCase , lowerCAmelCase , f=output_path.as_posix() , input_names=lowerCAmelCase , output_names=lowerCAmelCase , dynamic_axes=lowerCAmelCase , do_constant_folding=lowerCAmelCase , opset_version=lowerCAmelCase , ) @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = False ): '''simple docstring''' UpperCAmelCase = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): UpperCAmelCase = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: UpperCAmelCase = """cpu""" UpperCAmelCase = StableDiffusionPipeline.from_pretrained(lowerCAmelCase , torch_dtype=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase = Path(lowerCAmelCase ) # TEXT ENCODER UpperCAmelCase = pipeline.text_encoder.config.max_position_embeddings UpperCAmelCase = pipeline.text_encoder.config.hidden_size UpperCAmelCase = pipeline.tokenizer( """A sample prompt""" , padding="""max_length""" , max_length=pipeline.tokenizer.model_max_length , truncation=lowerCAmelCase , return_tensors="""pt""" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=lowerCAmelCase , dtype=torch.intaa )) , output_path=output_path / """text_encoder""" / """model.onnx""" , ordered_input_names=["""input_ids"""] , output_names=["""last_hidden_state""", """pooler_output"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase , ) del pipeline.text_encoder # UNET UpperCAmelCase = pipeline.unet.config.in_channels UpperCAmelCase = pipeline.unet.config.sample_size UpperCAmelCase = output_path / """unet""" / """model.onnx""" onnx_export( pipeline.unet , model_args=( torch.randn(2 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), torch.randn(2 ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), torch.randn(2 , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), False, ) , output_path=lowerCAmelCase , ordered_input_names=["""sample""", """timestep""", """encoder_hidden_states""", """return_dict"""] , output_names=["""out_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """timestep""": {0: """batch"""}, """encoder_hidden_states""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase , use_external_data_format=lowerCAmelCase , ) UpperCAmelCase = str(unet_path.absolute().as_posix() ) UpperCAmelCase = os.path.dirname(lowerCAmelCase ) UpperCAmelCase = onnx.load(lowerCAmelCase ) # clean up existing tensor files shutil.rmtree(lowerCAmelCase ) os.mkdir(lowerCAmelCase ) # collate external tensor files into one onnx.save_model( lowerCAmelCase , lowerCAmelCase , save_as_external_data=lowerCAmelCase , all_tensors_to_one_file=lowerCAmelCase , location="""weights.pb""" , convert_attribute=lowerCAmelCase , ) del pipeline.unet # VAE ENCODER UpperCAmelCase = pipeline.vae UpperCAmelCase = vae_encoder.config.in_channels UpperCAmelCase = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder UpperCAmelCase = lambda lowerCAmelCase , lowerCAmelCase : vae_encoder.encode(lowerCAmelCase , lowerCAmelCase )[0].sample() onnx_export( lowerCAmelCase , model_args=( torch.randn(1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), False, ) , output_path=output_path / """vae_encoder""" / """model.onnx""" , ordered_input_names=["""sample""", """return_dict"""] , output_names=["""latent_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase , ) # VAE DECODER UpperCAmelCase = pipeline.vae UpperCAmelCase = vae_decoder.config.latent_channels UpperCAmelCase = vae_decoder.config.out_channels # forward only through the decoder part UpperCAmelCase = vae_encoder.decode onnx_export( lowerCAmelCase , model_args=( torch.randn(1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: UpperCAmelCase = pipeline.safety_checker UpperCAmelCase = safety_checker.config.vision_config.num_channels UpperCAmelCase = safety_checker.config.vision_config.image_size UpperCAmelCase = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), torch.randn(1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), ) , output_path=output_path / """safety_checker""" / """model.onnx""" , ordered_input_names=["""clip_input""", """images"""] , output_names=["""out_images""", """has_nsfw_concepts"""] , dynamic_axes={ """clip_input""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """images""": {0: """batch""", 1: """height""", 2: """width""", 3: """channels"""}, } , opset=lowerCAmelCase , ) del pipeline.safety_checker UpperCAmelCase = OnnxRuntimeModel.from_pretrained(output_path / """safety_checker""" ) UpperCAmelCase = pipeline.feature_extractor else: UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_encoder""" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_decoder""" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / """text_encoder""" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / """unet""" ) , scheduler=pipeline.scheduler , safety_checker=lowerCAmelCase , feature_extractor=lowerCAmelCase , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(lowerCAmelCase ) print("""ONNX pipeline saved to""" , lowerCAmelCase ) del pipeline del onnx_pipeline UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(lowerCAmelCase , provider="""CPUExecutionProvider""" ) print("""ONNX pipeline is loadable""" ) if __name__ == "__main__": lowerCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_path''', type=str, required=True, help='''Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).''', ) parser.add_argument('''--output_path''', type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--opset''', default=1_4, type=int, help='''The version of the ONNX operator set to use.''', ) parser.add_argument('''--fp16''', action='''store_true''', default=False, help='''Export the models in `float16` mode''') lowerCAmelCase_ : Union[str, Any] = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
248
1
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class lowercase__( lowerCamelCase_ ): """simple docstring""" a :List[str] = '''autoformer''' a :Union[str, Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self : Tuple , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : str = "student_t" , SCREAMING_SNAKE_CASE_ : str = "nll" , SCREAMING_SNAKE_CASE_ : int = 1 , SCREAMING_SNAKE_CASE_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , SCREAMING_SNAKE_CASE_ : bool = True , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : int = 6_4 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 3_2 , SCREAMING_SNAKE_CASE_ : int = 3_2 , SCREAMING_SNAKE_CASE_ : str = "gelu" , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : float = 0.1 , SCREAMING_SNAKE_CASE_ : int = 1_0_0 , SCREAMING_SNAKE_CASE_ : float = 0.02 , SCREAMING_SNAKE_CASE_ : bool = True , SCREAMING_SNAKE_CASE_ : Dict=True , SCREAMING_SNAKE_CASE_ : int = 1_0 , SCREAMING_SNAKE_CASE_ : int = 2_5 , SCREAMING_SNAKE_CASE_ : int = 3 , **SCREAMING_SNAKE_CASE_ : Optional[int] , ) -> Any: # time series specific configuration lowercase_ = prediction_length lowercase_ = context_length if context_length is not None else prediction_length lowercase_ = distribution_output lowercase_ = loss lowercase_ = input_size lowercase_ = num_time_features lowercase_ = lags_sequence lowercase_ = scaling lowercase_ = num_dynamic_real_features lowercase_ = num_static_real_features lowercase_ = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(__snake_case ) != num_static_categorical_features: raise ValueError( '''The cardinality should be a list of the same length as `num_static_categorical_features`''' ) lowercase_ = cardinality else: lowercase_ = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(__snake_case ) != num_static_categorical_features: raise ValueError( '''The embedding dimension should be a list of the same length as `num_static_categorical_features`''' ) lowercase_ = embedding_dimension else: lowercase_ = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] lowercase_ = num_parallel_samples # Transformer architecture configuration lowercase_ = input_size * len(self.lags_sequence ) + self._number_of_features lowercase_ = d_model lowercase_ = encoder_attention_heads lowercase_ = decoder_attention_heads lowercase_ = encoder_ffn_dim lowercase_ = decoder_ffn_dim lowercase_ = encoder_layers lowercase_ = decoder_layers lowercase_ = dropout lowercase_ = attention_dropout lowercase_ = activation_dropout lowercase_ = encoder_layerdrop lowercase_ = decoder_layerdrop lowercase_ = activation_function lowercase_ = init_std lowercase_ = use_cache # Autoformer lowercase_ = label_length lowercase_ = moving_average lowercase_ = autocorrelation_factor super().__init__(is_encoder_decoder=__snake_case , **__snake_case ) @property def _lowercase ( self : int ) -> int: return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
30
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=lowerCamelCase_ ) class snake_case_ (lowerCamelCase_ ): UpperCAmelCase__ : str = field(default='''audio-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) UpperCAmelCase__ : ClassVar[Features] = Features({'''audio''': Audio()} ) UpperCAmelCase__ : ClassVar[Features] = Features({'''labels''': ClassLabel} ) UpperCAmelCase__ : str = "audio" UpperCAmelCase__ : str = "labels" def lowerCamelCase__( self :Optional[int] ,__snake_case :int ) -> str: if self.label_column not in features: raise ValueError(F'Column {self.label_column} is not present in features.' ) if not isinstance(features[self.label_column] ,__snake_case ): raise ValueError(F'Column {self.label_column} is not a ClassLabel.' ) a__ = copy.deepcopy(self ) a__ = self.label_schema.copy() a__ = features[self.label_column] a__ = label_schema return task_template @property def lowerCamelCase__( self :Dict ) -> Dict[str, str]: return { self.audio_column: "audio", self.label_column: "labels", }
240
0
'''simple docstring''' from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker lowerCAmelCase :Optional[Any] = '''CompVis/stable-diffusion-v1-1''' lowerCAmelCase :Any = '''CompVis/stable-diffusion-v1-2''' lowerCAmelCase :Optional[int] = '''CompVis/stable-diffusion-v1-3''' lowerCAmelCase :str = '''CompVis/stable-diffusion-v1-4''' class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self : Dict , _A : AutoencoderKL , _A : CLIPTextModel , _A : CLIPTokenizer , _A : UNetaDConditionModel , _A : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , _A : StableDiffusionSafetyChecker , _A : CLIPImageProcessor , _A : bool = True , ) -> Any: super()._init_() __magic_name__ : Any = StableDiffusionPipeline.from_pretrained(_A ) __magic_name__ : Tuple = StableDiffusionPipeline.from_pretrained(_A ) __magic_name__ : Dict = StableDiffusionPipeline.from_pretrained(_A ) __magic_name__ : str = StableDiffusionPipeline( vae=_A , text_encoder=_A , tokenizer=_A , unet=_A , scheduler=_A , safety_checker=_A , feature_extractor=_A , requires_safety_checker=_A , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def __lowerCAmelCase ( self : Union[str, Any] ) -> Dict[str, Any]: return {k: getattr(self , _A ) for k in self.config.keys() if not k.startswith('_' )} def __lowerCAmelCase ( self : Tuple , _A : Optional[Union[str, int]] = "auto" ) -> Any: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __magic_name__ : str = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_A ) def __lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: self.enable_attention_slicing(_A ) @torch.no_grad() def __lowerCAmelCase ( self : List[Any] , _A : Union[str, List[str]] , _A : int = 512 , _A : int = 512 , _A : int = 50 , _A : float = 7.5 , _A : Optional[Union[str, List[str]]] = None , _A : Optional[int] = 1 , _A : float = 0.0 , _A : Optional[torch.Generator] = None , _A : Optional[torch.FloatTensor] = None , _A : Optional[str] = "pil" , _A : bool = True , _A : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _A : int = 1 , **_A : Any , ) -> Any: return self.pipea( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) @torch.no_grad() def __lowerCAmelCase ( self : int , _A : Union[str, List[str]] , _A : int = 512 , _A : int = 512 , _A : int = 50 , _A : float = 7.5 , _A : Optional[Union[str, List[str]]] = None , _A : Optional[int] = 1 , _A : float = 0.0 , _A : Optional[torch.Generator] = None , _A : Optional[torch.FloatTensor] = None , _A : Optional[str] = "pil" , _A : bool = True , _A : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _A : int = 1 , **_A : Tuple , ) -> int: return self.pipea( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) @torch.no_grad() def __lowerCAmelCase ( self : Optional[int] , _A : Union[str, List[str]] , _A : int = 512 , _A : int = 512 , _A : int = 50 , _A : float = 7.5 , _A : Optional[Union[str, List[str]]] = None , _A : Optional[int] = 1 , _A : float = 0.0 , _A : Optional[torch.Generator] = None , _A : Optional[torch.FloatTensor] = None , _A : Optional[str] = "pil" , _A : bool = True , _A : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _A : int = 1 , **_A : Optional[int] , ) -> Union[str, Any]: return self.pipea( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) @torch.no_grad() def __lowerCAmelCase ( self : str , _A : Union[str, List[str]] , _A : int = 512 , _A : int = 512 , _A : int = 50 , _A : float = 7.5 , _A : Optional[Union[str, List[str]]] = None , _A : Optional[int] = 1 , _A : float = 0.0 , _A : Optional[torch.Generator] = None , _A : Optional[torch.FloatTensor] = None , _A : Optional[str] = "pil" , _A : bool = True , _A : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _A : int = 1 , **_A : List[str] , ) -> Optional[Any]: return self.pipea( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) @torch.no_grad() def __lowerCAmelCase ( self : Any , _A : Union[str, List[str]] , _A : int = 512 , _A : int = 512 , _A : int = 50 , _A : float = 7.5 , _A : Optional[Union[str, List[str]]] = None , _A : Optional[int] = 1 , _A : float = 0.0 , _A : Optional[torch.Generator] = None , _A : Optional[torch.FloatTensor] = None , _A : Optional[str] = "pil" , _A : bool = True , _A : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _A : int = 1 , **_A : Union[str, Any] , ) -> int: __magic_name__ : Optional[int] = 'cuda' if torch.cuda.is_available() else 'cpu' self.to(_A ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(F'`height` and `width` must be divisible by 8 but are {height} and {width}.' ) # Get first result from Stable Diffusion Checkpoint v1.1 __magic_name__ : int = self.textaimg_sda_a( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) # Get first result from Stable Diffusion Checkpoint v1.2 __magic_name__ : Dict = self.textaimg_sda_a( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) # Get first result from Stable Diffusion Checkpoint v1.3 __magic_name__ : List[Any] = self.textaimg_sda_a( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) # Get first result from Stable Diffusion Checkpoint v1.4 __magic_name__ : Optional[Any] = self.textaimg_sda_a( prompt=_A , height=_A , width=_A , num_inference_steps=_A , guidance_scale=_A , negative_prompt=_A , num_images_per_prompt=_A , eta=_A , generator=_A , latents=_A , output_type=_A , return_dict=_A , callback=_A , callback_steps=_A , **_A , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
275
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCamelCase ( lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' A_ : int = StableDiffusionXLImgaImgPipeline A_ : str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} A_ : Optional[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} A_ : Any = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A_ : Any = IMAGE_TO_IMAGE_IMAGE_PARAMS A_ : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS def __lowerCAmelCase ( self : Dict ) -> Optional[int]: torch.manual_seed(0 ) __magic_name__ : int = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , attention_head_dim=(2, 4) , use_linear_projection=_A , addition_embed_type='text_time' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __magic_name__ : str = EulerDiscreteScheduler( beta_start=0.0_0085 , beta_end=0.012 , steps_offset=1 , beta_schedule='scaled_linear' , timestep_spacing='leading' , ) torch.manual_seed(0 ) __magic_name__ : str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __magic_name__ : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=32 , ) __magic_name__ : Dict = CLIPTextModel(_A ) __magic_name__ : List[str] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=_A ) __magic_name__ : Optional[Any] = CLIPTextModelWithProjection(_A ) __magic_name__ : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=_A ) __magic_name__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'text_encoder_2': text_encoder_a, 'tokenizer_2': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def __lowerCAmelCase ( self : List[Any] , _A : List[str] , _A : Any=0 ) -> Union[str, Any]: __magic_name__ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A ) ).to(_A ) __magic_name__ : Dict = image / 2 + 0.5 if str(_A ).startswith('mps' ): __magic_name__ : Any = torch.manual_seed(_A ) else: __magic_name__ : int = torch.Generator(device=_A ).manual_seed(_A ) __magic_name__ : List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 5.0, 'output_type': 'numpy', 'strength': 0.75, } return inputs def __lowerCAmelCase ( self : Optional[Any] ) -> List[Any]: __magic_name__ : str = 'cpu' # ensure determinism for the device-dependent torch.Generator __magic_name__ : str = self.get_dummy_components() __magic_name__ : Any = StableDiffusionXLImgaImgPipeline(**_A ) __magic_name__ : List[Any] = sd_pipe.to(_A ) sd_pipe.set_progress_bar_config(disable=_A ) __magic_name__ : Any = self.get_dummy_inputs(_A ) __magic_name__ : Optional[int] = sd_pipe(**_A ).images __magic_name__ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __magic_name__ : Any = np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCAmelCase ( self : Optional[int] ) -> Optional[Any]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def __lowerCAmelCase ( self : List[Any] ) -> int: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def __lowerCAmelCase ( self : Any ) -> Union[str, Any]: pass def __lowerCAmelCase ( self : List[Any] ) -> Optional[int]: __magic_name__ : Dict = self.get_dummy_components() __magic_name__ : Optional[Any] = StableDiffusionXLImgaImgPipeline(**_A ) __magic_name__ : List[Any] = sd_pipe.to(_A ) __magic_name__ : str = sd_pipe.to(_A ) sd_pipe.set_progress_bar_config(disable=_A ) # forward without prompt embeds __magic_name__ : Union[str, Any] = self.get_dummy_inputs(_A ) __magic_name__ : Union[str, Any] = 3 * ['this is a negative prompt'] __magic_name__ : List[str] = negative_prompt __magic_name__ : int = 3 * [inputs['prompt']] __magic_name__ : Tuple = sd_pipe(**_A ) __magic_name__ : str = output.images[0, -3:, -3:, -1] # forward with prompt embeds __magic_name__ : Optional[Any] = self.get_dummy_inputs(_A ) __magic_name__ : Tuple = 3 * ['this is a negative prompt'] __magic_name__ : List[str] = 3 * [inputs.pop('prompt' )] ( ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ) : List[Any] = sd_pipe.encode_prompt(_A , negative_prompt=_A ) __magic_name__ : Tuple = sd_pipe( **_A , prompt_embeds=_A , negative_prompt_embeds=_A , pooled_prompt_embeds=_A , negative_pooled_prompt_embeds=_A , ) __magic_name__ : int = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : List[Any] ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : str , _A : Optional[int] , _A : Optional[Any]="cpu" , _A : List[str]=torch.floataa , _A : Any=0 ) -> str: __magic_name__ : List[str] = torch.Generator(device=_A ).manual_seed(_A ) __magic_name__ : Optional[Any] = np.random.RandomState(_A ).standard_normal((1, 4, 64, 64) ) __magic_name__ : Union[str, Any] = torch.from_numpy(_A ).to(device=_A , dtype=_A ) __magic_name__ : Optional[int] = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def __lowerCAmelCase ( self : Union[str, Any] ) -> Dict: __magic_name__ : str = DiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-base' ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __magic_name__ : Optional[int] = self.get_inputs(_A ) __magic_name__ : Union[str, Any] = pipe(**_A ).images __magic_name__ : Any = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __magic_name__ : List[Any] = np.array([0.4_9493, 0.4_7896, 0.4_0798, 0.5_4214, 0.5_3212, 0.4_8202, 0.4_7656, 0.4_6329, 0.4_8506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
275
1
"""simple docstring""" def _A (__a ) -> List[str]: """simple docstring""" return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _A (__a ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = number while duplicate > 0: SCREAMING_SNAKE_CASE_ : Tuple = divmod(lowerCAmelCase_ , 10 ) fact_sum += factorial(lowerCAmelCase_ ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") UpperCAmelCase_ : int = int(input("""Enter number: """).strip()) print( f'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
91
from __future__ import annotations def snake_case_ ( lowerCAmelCase_ : int , lowerCAmelCase_ : int ): if partitions <= 0: raise ValueError("""partitions must be a positive number!""" ) if partitions > number_of_bytes: raise ValueError("""partitions can not > number_of_bytes!""" ) __lowercase : Dict = number_of_bytes // partitions __lowercase : Union[str, Any] = [] for i in range(lowerCAmelCase_ ): __lowercase : str = i * bytes_per_partition + 1 __lowercase : List[Any] = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F"{start_bytes}-{end_bytes}" ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
233
0
from __future__ import annotations from PIL import Image # Define glider example __lowerCamelCase : Optional[Any] = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example __lowerCamelCase : List[Any] = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : list[list[int]] ) -> list[list[int]]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] for i in range(len(__UpperCamelCase ) ): SCREAMING_SNAKE_CASE__ = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours SCREAMING_SNAKE_CASE__ = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(__UpperCamelCase ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(__UpperCamelCase ) - 1: neighbour_count += cells[i + 1][j] if i < len(__UpperCamelCase ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. SCREAMING_SNAKE_CASE__ = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(__UpperCamelCase ) return next_generation def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int ) -> list[Image.Image]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] for _ in range(__UpperCamelCase ): # Create output image SCREAMING_SNAKE_CASE__ = Image.new("""RGB""" , (len(cells[0] ), len(__UpperCamelCase )) ) SCREAMING_SNAKE_CASE__ = img.load() # Save cells to image for x in range(len(__UpperCamelCase ) ): for y in range(len(cells[0] ) ): SCREAMING_SNAKE_CASE__ = 2_55 - cells[y][x] * 2_55 SCREAMING_SNAKE_CASE__ = (colour, colour, colour) # Save image images.append(__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = new_generation(__UpperCamelCase ) return images if __name__ == "__main__": __lowerCamelCase : Any = generate_images(GLIDER, 16) images[0].save('''out.gif''', save_all=True, append_images=images[1:])
368
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch __lowerCamelCase : int = logging.get_logger(__name__) class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = ["pixel_values"] def __init__( self : int , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = PILImageResampling.BILINEAR , _lowercase : bool = True , _lowercase : Union[int, float] = 1 / 2_55 , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : bool = True , **_lowercase : List[Any] , ): """simple docstring""" super().__init__(**_lowercase ) SCREAMING_SNAKE_CASE__ = size if size is not None else {"""shortest_edge""": 2_24} SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else {"""height""": 2_56, """width""": 2_56} SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , param_name="""crop_size""" ) SCREAMING_SNAKE_CASE__ = do_resize SCREAMING_SNAKE_CASE__ = size SCREAMING_SNAKE_CASE__ = resample SCREAMING_SNAKE_CASE__ = do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor SCREAMING_SNAKE_CASE__ = do_center_crop SCREAMING_SNAKE_CASE__ = crop_size SCREAMING_SNAKE_CASE__ = do_flip_channel_order def __a ( self : List[Any] , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : PILImageResampling = PIL.Image.BILINEAR , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : str , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) SCREAMING_SNAKE_CASE__ = get_resize_output_image_size(_lowercase , size=size["""shortest_edge"""] , default_to_square=_lowercase ) return resize(_lowercase , size=_lowercase , resample=_lowercase , data_format=_lowercase , **_lowercase ) def __a ( self : str , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Any , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_lowercase , size=(size["""height"""], size["""width"""]) , data_format=_lowercase , **_lowercase ) def __a ( self : Optional[Any] , _lowercase : np.ndarray , _lowercase : Union[int, float] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Any , ): """simple docstring""" return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def __a ( self : Tuple , _lowercase : np.ndarray , _lowercase : Optional[Union[str, ChannelDimension]] = None ): """simple docstring""" return flip_channel_order(_lowercase , data_format=_lowercase ) def __a ( self : List[str] , _lowercase : ImageInput , _lowercase : bool = None , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = None , _lowercase : bool = None , _lowercase : float = None , _lowercase : bool = None , _lowercase : Dict[str, int] = None , _lowercase : bool = None , _lowercase : Optional[Union[str, TensorType]] = None , _lowercase : ChannelDimension = ChannelDimension.FIRST , **_lowercase : int , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE__ = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE__ = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE__ = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE__ = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) SCREAMING_SNAKE_CASE__ = size if size is not None else self.size SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , param_name="""crop_size""" ) SCREAMING_SNAKE_CASE__ = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE__ = [to_numpy_array(_lowercase ) for image in images] if do_resize: SCREAMING_SNAKE_CASE__ = [self.resize(image=_lowercase , size=_lowercase , resample=_lowercase ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE__ = [self.center_crop(image=_lowercase , size=_lowercase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE__ = [self.rescale(image=_lowercase , scale=_lowercase ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: SCREAMING_SNAKE_CASE__ = [self.flip_channel_order(image=_lowercase ) for image in images] SCREAMING_SNAKE_CASE__ = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] SCREAMING_SNAKE_CASE__ = {"""pixel_values""": images} return BatchFeature(data=_lowercase , tensor_type=_lowercase ) def __a ( self : List[Any] , _lowercase : Dict , _lowercase : List[Tuple] = None ): """simple docstring""" SCREAMING_SNAKE_CASE__ = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowercase ) != len(_lowercase ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(_lowercase ): SCREAMING_SNAKE_CASE__ = target_sizes.numpy() SCREAMING_SNAKE_CASE__ = [] for idx in range(len(_lowercase ) ): SCREAMING_SNAKE_CASE__ = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=_lowercase ) SCREAMING_SNAKE_CASE__ = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowercase ) else: SCREAMING_SNAKE_CASE__ = logits.argmax(dim=1 ) SCREAMING_SNAKE_CASE__ = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
204
0
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class __a ( nn.Module ): def __init__( self , lowerCAmelCase__ = 16 , lowerCAmelCase__ = 88 , lowerCAmelCase__ = None , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 32 , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = "geglu" , lowerCAmelCase__ = None , ) -> List[str]: '''simple docstring''' super().__init__() lowercase__: Optional[Any] = nn.ModuleList( [ TransformeraDModel( num_attention_heads=__snake_case , attention_head_dim=__snake_case , in_channels=__snake_case , num_layers=__snake_case , dropout=__snake_case , norm_num_groups=__snake_case , cross_attention_dim=__snake_case , attention_bias=__snake_case , sample_size=__snake_case , num_vector_embeds=__snake_case , activation_fn=__snake_case , num_embeds_ada_norm=__snake_case , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference lowercase__: Union[str, Any] = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` lowercase__: Optional[Any] = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` lowercase__: Optional[Any] = [1, 0] def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__ = True , ) -> Optional[int]: '''simple docstring''' lowercase__: int = hidden_states lowercase__: Dict = [] lowercase__: List[str] = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens lowercase__: Dict = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] lowercase__: Dict = self.transformer_index_for_condition[i] lowercase__: Tuple = self.transformers[transformer_index]( __snake_case , encoder_hidden_states=__snake_case , timestep=__snake_case , cross_attention_kwargs=__snake_case , return_dict=__snake_case , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] lowercase__: Any = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) lowercase__: Union[str, Any] = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=__snake_case )
196
import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _lowerCAmelCase : Optional[Any] = False class __magic_name__ ( unittest.TestCase ): pass @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase ): def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) __a =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) __a =torch.manual_seed(0 ) __a =pipe( image=__snake_case , generator=__snake_case , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' , ).images __a =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __a =np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
218
0
def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : list[float] ): if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError('Cash flows list cannot be empty' ) a__ = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(__lowerCAmelCase ) ) return round(__lowerCAmelCase , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
109
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() snake_case : str = logging.get_logger(__name__) snake_case : Optional[Any] = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn.grep_linear''': '''encoder.layers.*.attention.gru_rel_pos_linear''', '''self_attn.relative_attention_bias''': '''encoder.layers.*.attention.rel_attn_embed''', '''self_attn.grep_a''': '''encoder.layers.*.attention.gru_rel_pos_const''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''ctc_proj''', '''mask_emb''': '''masked_spec_embed''', } snake_case : Dict = [ '''ctc_proj''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def __lowercase ( __lowerCAmelCase : Dict , __lowerCAmelCase : List[str] , __lowerCAmelCase : str , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Dict ): for attribute in key.split('.' ): a__ = getattr(__lowerCAmelCase , __lowerCAmelCase ) if weight_type is not None: a__ = getattr(__lowerCAmelCase , __lowerCAmelCase ).shape else: a__ = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__ = value elif weight_type == "weight_g": a__ = value elif weight_type == "weight_v": a__ = value elif weight_type == "bias": a__ = value else: a__ = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __lowercase ( __lowerCAmelCase : List[str] , __lowerCAmelCase : List[Any] ): a__ = [] a__ = fairseq_model.state_dict() a__ = hf_model.feature_extractor for name, value in fairseq_dict.items(): a__ = False if "conv_layers" in name: load_conv_layer( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) a__ = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: a__ = True if "*" in mapped_key: a__ = name.split(__lowerCAmelCase )[0].split('.' )[-2] a__ = mapped_key.replace('*' , __lowerCAmelCase ) if "weight_g" in name: a__ = 'weight_g' elif "weight_v" in name: a__ = 'weight_v' elif "bias" in name and "relative_attention_bias" not in name: a__ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj a__ = 'weight' else: a__ = None set_recursively(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) continue if not is_used: unused_weights.append(__lowerCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __lowercase ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Dict , __lowerCAmelCase : int , __lowerCAmelCase : Optional[int] ): a__ = full_name.split('conv_layers.' )[-1] a__ = name.split('.' ) a__ = int(items[0] ) a__ = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) a__ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) a__ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) a__ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) a__ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__lowerCAmelCase ) @torch.no_grad() def __lowercase ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any , __lowerCAmelCase : List[Any]=None ): # load the pre-trained checkpoints a__ = torch.load(__lowerCAmelCase ) a__ = WavLMConfigOrig(checkpoint['cfg'] ) a__ = WavLMOrig(__lowerCAmelCase ) model.load_state_dict(checkpoint['model'] ) model.eval() if config_path is not None: a__ = WavLMConfig.from_pretrained(__lowerCAmelCase ) else: a__ = WavLMConfig() a__ = WavLMModel(__lowerCAmelCase ) recursively_load_weights(__lowerCAmelCase , __lowerCAmelCase ) hf_wavlm.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": snake_case : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') snake_case : Union[str, Any] = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
109
1
"""simple docstring""" def _snake_case ( _snake_case : Dict = 10 , _snake_case : List[Any] = 10_00 , _snake_case : Dict = True ) -> Any: '''simple docstring''' assert ( isinstance(_snake_case , _snake_case ) and isinstance(_snake_case , _snake_case ) and isinstance(_snake_case , _snake_case ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError('Invalid value for min_val or max_val (min_value < max_value)' ) return min_val if option else max_val def _snake_case ( _snake_case : Union[str, Any] , _snake_case : List[str] ) -> Tuple: '''simple docstring''' return int((number_a + number_a) / 2 ) def _snake_case ( _snake_case : Optional[Any] , _snake_case : Optional[Any] , _snake_case : Tuple ) -> Dict: '''simple docstring''' assert ( isinstance(_snake_case , _snake_case ) and isinstance(_snake_case , _snake_case ) and isinstance(_snake_case , _snake_case ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError('argument value for lower and higher must be(lower > higher)' ) if not lower < to_guess < higher: raise ValueError( 'guess value must be within the range of lower and higher value' ) def answer(_snake_case : List[str] ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print('started...' ) _A = lower _A = higher _A = [] while True: _A = get_avg(_snake_case , _snake_case ) last_numbers.append(_snake_case ) if answer(_snake_case ) == "low": _A = number elif answer(_snake_case ) == "high": _A = number else: break print(F'''guess the number : {last_numbers[-1]}''' ) print(F'''details : {last_numbers!s}''' ) def _snake_case ( ) -> str: '''simple docstring''' _A = int(input('Enter lower value : ' ).strip() ) _A = int(input('Enter high value : ' ).strip() ) _A = int(input('Enter value to guess : ' ).strip() ) guess_the_number(_snake_case , _snake_case , _snake_case ) if __name__ == "__main__": main()
315
"""simple docstring""" import math import sys def A__ ( UpperCamelCase ): A = "" try: with open(UpperCamelCase , "rb" ) as binary_file: A = binary_file.read() for dat in data: A = F"{dat:08b}" result += curr_byte return result except OSError: print("File not accessible" ) sys.exit() def A__ ( UpperCamelCase ): A = {"0": "0", "1": "1"} A, A = "", "" A = len(UpperCamelCase ) for i in range(len(UpperCamelCase ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue A = lexicon[curr_string] result += last_match_id A = last_match_id + "0" if math.loga(UpperCamelCase ).is_integer(): A = {} for curr_key in list(UpperCamelCase ): A = lexicon.pop(UpperCamelCase ) A = new_lex A = last_match_id + "1" index += 1 A = "" return result def A__ ( UpperCamelCase , UpperCamelCase ): A = 8 try: with open(UpperCamelCase , "wb" ) as opened_file: A = [ to_write[i : i + byte_length] for i in range(0 , len(UpperCamelCase ) , UpperCamelCase ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append("10000000" ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(UpperCamelCase , 2 ).to_bytes(1 , byteorder="big" ) ) except OSError: print("File not accessible" ) sys.exit() def A__ ( UpperCamelCase ): A = 0 for letter in data_bits: if letter == "1": break counter += 1 A = data_bits[counter:] A = data_bits[counter + 1 :] return data_bits def A__ ( UpperCamelCase , UpperCamelCase ): A = read_file_binary(UpperCamelCase ) A = remove_prefix(UpperCamelCase ) A = decompress_data(UpperCamelCase ) write_file_binary(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
292
0
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __snake_case = logging.get_logger(__name__) def A_ ( _lowerCAmelCase : List[Any], _lowerCAmelCase : List[Any]=False ): """simple docstring""" _a = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith('''head''' ): _a = "segformer.encoder." + key if key.startswith('''backbone''' ): _a = key.replace('''backbone''', '''segformer.encoder''' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 _a = key[key.find('''patch_embed''' ) + len('''patch_embed''' )] _a = key.replace(f'patch_embed{idx}', f'patch_embeddings.{int(_lowerCAmelCase )-1}' ) if "norm" in key: _a = key.replace('''norm''', '''layer_norm''' ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 _a = key[key.find('''segformer.encoder.layer_norm''' ) + len('''segformer.encoder.layer_norm''' )] _a = key.replace(f'layer_norm{idx}', f'layer_norm.{int(_lowerCAmelCase )-1}' ) if "layer_norm1" in key: _a = key.replace('''layer_norm1''', '''layer_norm_1''' ) if "layer_norm2" in key: _a = key.replace('''layer_norm2''', '''layer_norm_2''' ) if "block" in key: # replace for example block1 by block.0 _a = key[key.find('''block''' ) + len('''block''' )] _a = key.replace(f'block{idx}', f'block.{int(_lowerCAmelCase )-1}' ) if "attn.q" in key: _a = key.replace('''attn.q''', '''attention.self.query''' ) if "attn.proj" in key: _a = key.replace('''attn.proj''', '''attention.output.dense''' ) if "attn" in key: _a = key.replace('''attn''', '''attention.self''' ) if "fc1" in key: _a = key.replace('''fc1''', '''dense1''' ) if "fc2" in key: _a = key.replace('''fc2''', '''dense2''' ) if "linear_pred" in key: _a = key.replace('''linear_pred''', '''classifier''' ) if "linear_fuse" in key: _a = key.replace('''linear_fuse.conv''', '''linear_fuse''' ) _a = key.replace('''linear_fuse.bn''', '''batch_norm''' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 _a = key[key.find('''linear_c''' ) + len('''linear_c''' )] _a = key.replace(f'linear_c{idx}', f'linear_c.{int(_lowerCAmelCase )-1}' ) if key.startswith('''head''' ): _a = key.replace('''head''', '''classifier''' ) _a = value return new_state_dict def A_ ( _lowerCAmelCase : List[Any], _lowerCAmelCase : Any ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) _a = state_dict.pop(f'segformer.encoder.block.{i}.{j}.attention.self.kv.weight' ) _a = state_dict.pop(f'segformer.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict _a = kv_weight[ : config.hidden_sizes[i], : ] _a = kv_bias[: config.hidden_sizes[i]] _a = kv_weight[ config.hidden_sizes[i] :, : ] _a = kv_bias[ config.hidden_sizes[i] : ] def A_ ( ): """simple docstring""" _a = "http://images.cocodataset.org/val2017/000000039769.jpg" _a = Image.open(requests.get(_lowerCAmelCase, stream=_lowerCAmelCase ).raw ) return image @torch.no_grad() def A_ ( _lowerCAmelCase : List[str], _lowerCAmelCase : int, _lowerCAmelCase : str ): """simple docstring""" _a = SegformerConfig() _a = False # set attributes based on model_name _a = "huggingface/label-files" if "segformer" in model_name: _a = model_name[len('''segformer.''' ) : len('''segformer.''' ) + 2] if "ade" in model_name: _a = 1_50 _a = "ade20k-id2label.json" _a = (1, 1_50, 1_28, 1_28) elif "city" in model_name: _a = 19 _a = "cityscapes-id2label.json" _a = (1, 19, 1_28, 1_28) else: raise ValueError(f'Model {model_name} not supported' ) elif "mit" in model_name: _a = True _a = model_name[4:6] _a = 10_00 _a = "imagenet-1k-id2label.json" _a = (1, 10_00) else: raise ValueError(f'Model {model_name} not supported' ) # set config attributes _a = json.load(open(hf_hub_download(_lowerCAmelCase, _lowerCAmelCase, repo_type='''dataset''' ), '''r''' ) ) _a = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} _a = idalabel _a = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": _a = [64, 1_28, 3_20, 5_12] _a = 2_56 elif size == "b2": _a = [64, 1_28, 3_20, 5_12] _a = 7_68 _a = [3, 4, 6, 3] elif size == "b3": _a = [64, 1_28, 3_20, 5_12] _a = 7_68 _a = [3, 4, 18, 3] elif size == "b4": _a = [64, 1_28, 3_20, 5_12] _a = 7_68 _a = [3, 8, 27, 3] elif size == "b5": _a = [64, 1_28, 3_20, 5_12] _a = 7_68 _a = [3, 6, 40, 3] else: raise ValueError(f'Size {size} not supported' ) # load image processor (only resize + normalize) _a = SegformerImageProcessor( image_scale=(5_12, 5_12), keep_ratio=_lowerCAmelCase, align=_lowerCAmelCase, do_random_crop=_lowerCAmelCase ) # prepare image _a = prepare_img() _a = image_processor(images=_lowerCAmelCase, return_tensors='''pt''' ).pixel_values logger.info(f'Converting model {model_name}...' ) # load original state dict if encoder_only: _a = torch.load(_lowerCAmelCase, map_location=torch.device('''cpu''' ) ) else: _a = torch.load(_lowerCAmelCase, map_location=torch.device('''cpu''' ) )["state_dict"] # rename keys _a = rename_keys(_lowerCAmelCase, encoder_only=_lowerCAmelCase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(_lowerCAmelCase, _lowerCAmelCase ) # create HuggingFace model and load state dict if encoder_only: _a = False _a = SegformerForImageClassification(_lowerCAmelCase ) else: _a = SegformerForSemanticSegmentation(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() # forward pass _a = model(_lowerCAmelCase ) _a = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": _a = torch.tensor( [ [[-4.6_3_1_0, -5.5_2_3_2, -6.2_3_5_6], [-5.1_9_2_1, -6.1_4_4_4, -6.5_9_9_6], [-5.4_4_2_4, -6.2_7_9_0, -6.7_5_7_4]], [[-1_2.1_3_9_1, -1_3.3_1_2_2, -1_3.9_5_5_4], [-1_2.8_7_3_2, -1_3.9_3_5_2, -1_4.3_5_6_3], [-1_2.9_4_3_8, -1_3.8_2_2_6, -1_4.2_5_1_3]], [[-1_2.5_1_3_4, -1_3.4_6_8_6, -1_4.4_9_1_5], [-1_2.8_6_6_9, -1_4.4_3_4_3, -1_4.7_7_5_8], [-1_3.2_5_2_3, -1_4.5_8_1_9, -1_5.0_6_9_4]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": _a = torch.tensor( [ [[-7.5_8_2_0, -8.7_2_3_1, -8.3_2_1_5], [-8.0_6_0_0, -1_0.3_5_2_9, -1_0.0_3_0_4], [-7.5_2_0_8, -9.4_1_0_3, -9.6_2_3_9]], [[-1_2.6_9_1_8, -1_3.8_9_9_4, -1_3.7_1_3_7], [-1_3.3_1_9_6, -1_5.7_5_2_3, -1_5.4_7_8_9], [-1_2.9_3_4_3, -1_4.8_7_5_7, -1_4.9_6_8_9]], [[-1_1.1_9_1_1, -1_1.9_4_2_1, -1_1.3_2_4_3], [-1_1.3_3_4_2, -1_3.6_8_3_9, -1_3.3_5_8_1], [-1_0.3_9_0_9, -1_2.1_8_3_2, -1_2.4_8_5_8]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": _a = torch.tensor( [ [[-1_1.8_1_7_3, -1_4.3_8_5_0, -1_6.3_1_2_8], [-1_4.5_6_4_8, -1_6.5_8_0_4, -1_8.6_5_6_8], [-1_4.7_2_2_3, -1_5.7_3_8_7, -1_8.4_2_1_8]], [[-1_5.7_2_9_0, -1_7.9_1_7_1, -1_9.4_4_2_3], [-1_8.3_1_0_5, -1_9.9_4_4_8, -2_1.4_6_6_1], [-1_7.9_2_9_6, -1_8.6_4_9_7, -2_0.7_9_1_0]], [[-1_5.0_7_8_3, -1_7.0_3_3_6, -1_8.2_7_8_9], [-1_6.8_7_7_1, -1_8.6_8_7_0, -2_0.1_6_1_2], [-1_6.2_4_5_4, -1_7.1_4_2_6, -1_9.5_0_5_5]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": _a = torch.tensor( [ [[-9.0_8_7_8, -1_0.2_0_8_1, -1_0.1_8_9_1], [-9.3_1_4_4, -1_0.7_9_4_1, -1_0.9_8_4_3], [-9.2_2_9_4, -1_0.3_8_5_5, -1_0.5_7_0_4]], [[-1_2.2_3_1_6, -1_3.9_0_6_8, -1_3.6_1_0_2], [-1_2.9_1_6_1, -1_4.3_7_0_2, -1_4.3_2_3_5], [-1_2.5_2_3_3, -1_3.7_1_7_4, -1_3.7_9_3_2]], [[-1_4.6_2_7_5, -1_5.2_4_9_0, -1_4.9_7_2_7], [-1_4.3_4_0_0, -1_5.9_6_8_7, -1_6.2_8_2_7], [-1_4.1_4_8_4, -1_5.4_0_3_3, -1_5.8_9_3_7]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": _a = torch.tensor( [ [[-1_2.3_1_4_4, -1_3.2_4_4_7, -1_4.0_8_0_2], [-1_3.3_6_1_4, -1_4.5_8_1_6, -1_5.6_1_1_7], [-1_3.3_3_4_0, -1_4.4_4_3_3, -1_6.2_2_1_9]], [[-1_9.2_7_8_1, -2_0.4_1_2_8, -2_0.7_5_0_6], [-2_0.6_1_5_3, -2_1.6_5_6_6, -2_2.0_9_9_8], [-1_9.9_8_0_0, -2_1.0_4_3_0, -2_2.1_4_9_4]], [[-1_8.8_7_3_9, -1_9.7_8_0_4, -2_1.1_8_3_4], [-2_0.1_2_3_3, -2_1.6_7_6_5, -2_3.2_9_4_4], [-2_0.0_3_1_5, -2_1.2_6_4_1, -2_3.6_9_4_4]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": _a = torch.tensor( [ [[-9.5_5_2_4, -1_2.0_8_3_5, -1_1.7_3_4_8], [-1_0.5_2_2_9, -1_3.6_4_4_6, -1_4.5_6_6_2], [-9.5_8_4_2, -1_2.8_8_5_1, -1_3.9_4_1_4]], [[-1_5.3_4_3_2, -1_7.5_3_2_3, -1_7.0_8_1_8], [-1_6.3_3_3_0, -1_8.9_2_5_5, -1_9.2_1_0_1], [-1_5.1_3_4_0, -1_7.7_8_4_8, -1_8.3_9_7_1]], [[-1_2.6_0_7_2, -1_4.9_4_8_6, -1_4.6_6_3_1], [-1_3.7_6_2_9, -1_7.0_9_0_7, -1_7.7_7_4_5], [-1_2.7_8_9_9, -1_6.1_6_9_5, -1_7.1_6_7_1]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": _a = torch.tensor( [ [[-1_1.9_2_9_5, -1_3.4_0_5_7, -1_4.8_1_0_6], [-1_3.3_4_3_1, -1_4.8_1_7_9, -1_5.3_7_8_1], [-1_4.2_8_3_6, -1_5.5_9_4_2, -1_6.1_5_8_8]], [[-1_1.4_9_0_6, -1_2.8_0_6_7, -1_3.6_5_6_4], [-1_3.1_1_8_9, -1_4.0_5_0_0, -1_4.1_5_4_3], [-1_3.8_7_4_8, -1_4.5_1_3_6, -1_4.8_7_8_9]], [[0.5_3_7_4, 0.1_0_6_7, -0.4_7_4_2], [0.1_1_4_1, -0.2_2_5_5, -0.7_0_9_9], [-0.3_0_0_0, -0.5_9_2_4, -1.3_1_0_5]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": _a = torch.tensor( [ [[-7.8_2_1_7, -9.8_7_6_7, -1_0.1_7_1_7], [-9.4_4_3_8, -1_0.9_0_5_8, -1_1.4_0_4_7], [-9.7_9_3_9, -1_2.3_4_9_5, -1_2.1_0_7_9]], [[-7.1_5_1_4, -9.5_3_3_6, -1_0.0_8_6_0], [-9.7_7_7_6, -1_1.6_8_2_2, -1_1.8_4_3_9], [-1_0.1_4_1_1, -1_2.7_6_5_5, -1_2.8_9_7_2]], [[0.3_0_2_1, 0.0_8_0_5, -0.2_3_1_0], [-0.0_3_2_8, -0.1_6_0_5, -0.2_7_1_4], [-0.1_4_0_8, -0.5_4_7_7, -0.6_9_7_6]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": _a = torch.tensor( [ [ [-1.1_3_7_2e0_1, -1.2_7_8_7e0_1, -1.3_4_7_7e0_1], [-1.2_5_3_6e0_1, -1.4_1_9_4e0_1, -1.4_4_0_9e0_1], [-1.3_2_1_7e0_1, -1.4_8_8_8e0_1, -1.5_3_2_7e0_1], ], [ [-1.4_7_9_1e0_1, -1.7_1_2_2e0_1, -1.8_2_7_7e0_1], [-1.7_1_6_3e0_1, -1.9_1_9_2e0_1, -1.9_5_3_3e0_1], [-1.7_8_9_7e0_1, -1.9_9_9_1e0_1, -2.0_3_1_5e0_1], ], [ [7.6_7_2_3e-0_1, 4.1_9_2_1e-0_1, -7.7_8_7_8e-0_2], [4.7_7_7_2e-0_1, 9.5_5_5_7e-0_3, -2.8_0_8_2e-0_1], [3.6_0_3_2e-0_1, -2.4_8_2_6e-0_1, -5.1_1_6_8e-0_1], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": _a = torch.tensor( [ [[-9.4_9_5_9, -1_1.3_0_8_7, -1_1.7_4_7_9], [-1_1.0_0_2_5, -1_2.6_5_4_0, -1_2.3_3_1_9], [-1_1.4_0_6_4, -1_3.0_4_8_7, -1_2.9_9_0_5]], [[-9.8_9_0_5, -1_1.3_0_8_4, -1_2.0_8_5_4], [-1_1.1_7_2_6, -1_2.7_6_9_8, -1_2.9_5_8_3], [-1_1.5_9_8_5, -1_3.3_2_7_8, -1_4.1_7_7_4]], [[0.2_2_1_3, 0.0_1_9_2, -0.2_4_6_6], [-0.1_7_3_1, -0.4_2_1_3, -0.4_8_7_4], [-0.3_1_2_6, -0.6_5_4_1, -1.1_3_8_9]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": _a = torch.tensor( [ [[-1_3.5_7_4_8, -1_3.9_1_1_1, -1_2.6_5_0_0], [-1_4.3_5_0_0, -1_5.3_6_8_3, -1_4.2_3_2_8], [-1_4.7_5_3_2, -1_6.0_4_2_4, -1_5.6_0_8_7]], [[-1_7.1_6_5_1, -1_5.8_7_2_5, -1_2.9_6_5_3], [-1_7.2_5_8_0, -1_7.3_7_1_8, -1_4.8_2_2_3], [-1_6.6_0_5_8, -1_6.8_7_8_3, -1_6.7_4_5_2]], [[-3.6_4_5_6, -3.0_2_0_9, -1.4_2_0_3], [-3.0_7_9_7, -3.1_9_5_9, -2.0_0_0_0], [-1.8_7_5_7, -1.9_2_1_7, -1.6_9_9_7]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": _a = torch.tensor( [ [[-1_6.0_9_7_6, -1_6.4_8_5_6, -1_7.3_9_6_2], [-1_6.6_2_3_4, -1_9.0_3_4_2, -1_9.7_6_8_5], [-1_6.0_9_0_0, -1_8.0_6_6_1, -1_9.1_1_8_0]], [[-1_8.4_7_5_0, -1_8.8_4_8_8, -1_9.5_0_7_4], [-1_9.4_0_3_0, -2_2.1_5_7_0, -2_2.5_9_7_7], [-1_9.1_1_9_1, -2_0.8_4_8_6, -2_2.3_7_8_3]], [[-4.5_1_7_8, -5.5_0_3_7, -6.5_1_0_9], [-5.0_8_8_4, -7.2_1_7_4, -8.0_3_3_4], [-4.4_1_5_6, -5.8_1_1_7, -7.2_9_7_0]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": _a = torch.tensor( [ [[-1_4.2_0_8_1, -1_4.4_7_3_2, -1_4.1_9_7_7], [-1_4.5_8_6_7, -1_6.4_4_2_3, -1_6.6_3_5_6], [-1_3.4_4_4_1, -1_4.9_6_8_5, -1_6.8_6_9_6]], [[-1_4.4_5_7_6, -1_4.7_0_7_3, -1_5.0_4_5_1], [-1_5.0_8_1_6, -1_7.6_2_3_7, -1_7.9_8_7_3], [-1_4.4_2_1_3, -1_6.0_1_9_9, -1_8.5_9_9_2]], [[-4.7_3_4_9, -4.9_5_8_8, -5.0_9_6_6], [-4.3_2_1_0, -6.9_3_2_5, -7.2_5_9_1], [-3.4_3_1_2, -4.7_4_8_4, -7.1_9_1_7]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": _a = torch.tensor( [ [[-1_1.7_7_3_7, -1_1.9_5_2_6, -1_1.3_2_7_3], [-1_3.6_6_9_2, -1_4.4_5_7_4, -1_3.8_8_7_8], [-1_3.8_9_3_7, -1_4.6_9_2_4, -1_5.9_3_4_5]], [[-1_4.6_7_0_6, -1_4.5_3_3_0, -1_4.1_3_0_6], [-1_6.1_5_0_2, -1_6.8_1_8_0, -1_6.4_2_6_9], [-1_6.8_3_3_8, -1_7.8_9_3_9, -2_0.1_7_4_6]], [[1.0_4_9_1, 0.8_2_8_9, 1.0_3_1_0], [1.1_0_4_4, 0.5_2_1_9, 0.8_0_5_5], [1.0_8_9_9, 0.6_9_2_6, 0.5_5_9_0]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": _a = torch.tensor( [ [[-1_2.5_6_4_1, -1_3.4_7_7_7, -1_3.0_6_8_4], [-1_3.9_5_8_7, -1_5.8_9_8_3, -1_6.6_5_5_7], [-1_3.3_1_0_9, -1_5.7_3_5_0, -1_6.3_1_4_1]], [[-1_4.7_0_7_4, -1_5.4_3_5_2, -1_4.5_9_4_4], [-1_6.6_3_5_3, -1_8.1_6_6_3, -1_8.6_1_2_0], [-1_5.1_7_0_2, -1_8.0_3_2_9, -1_8.1_5_4_7]], [[-1.7_9_9_0, -2.0_9_5_1, -1.7_7_8_4], [-2.6_3_9_7, -3.8_2_4_5, -3.9_6_8_6], [-1.5_2_6_4, -2.8_1_2_6, -2.9_3_1_6]], ] ) else: _a = logits.argmax(-1 ).item() print('''Predicted class:''', model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3], _lowerCAmelCase, atol=1e-2 ) # finally, save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) image_processor.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __snake_case = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
354
"""simple docstring""" import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor __snake_case = logging.get_logger(__name__) class __lowerCamelCase ( a__ ): '''simple docstring''' def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> None: warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
153
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer a : Any = logging.get_logger(__name__) a : List[str] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} a : Dict = { 'vocab_file': {'mobilebert-uncased': 'https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt'}, 'tokenizer_file': { 'mobilebert-uncased': 'https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json' }, } a : Tuple = {'mobilebert-uncased': 512} a : Optional[int] = {} class a ( _lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_INIT_CONFIGURATION snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = MobileBertTokenizer def __init__( self : int , lowercase_ : Tuple=None , lowercase_ : int=None , lowercase_ : int=True , lowercase_ : int="[UNK]" , lowercase_ : Union[str, Any]="[SEP]" , lowercase_ : str="[PAD]" , lowercase_ : List[str]="[CLS]" , lowercase_ : List[str]="[MASK]" , lowercase_ : Union[str, Any]=True , lowercase_ : List[str]=None , **lowercase_ : int , ): super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , tokenize_chinese_chars=lowercase_ , strip_accents=lowercase_ , **lowercase_ , ) snake_case_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowercase_ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowercase_ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowercase_ ) != tokenize_chinese_chars ): snake_case_ = getattr(lowercase_ , normalizer_state.pop('''type''' ) ) snake_case_ = do_lower_case snake_case_ = strip_accents snake_case_ = tokenize_chinese_chars snake_case_ = normalizer_class(**lowercase_ ) snake_case_ = do_lower_case def A_ ( self : str , lowercase_ : List[str] , lowercase_ : Optional[int]=None ): snake_case_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ): snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Union[str, Any] , lowercase_ : str , lowercase_ : Optional[str] = None ): snake_case_ = self._tokenizer.model.save(lowercase_ , name=lowercase_ ) return tuple(lowercase_ )
56
"""simple docstring""" import os def snake_case ( ): with open(os.path.dirname(A__ ) + "/grid.txt" ) as f: UpperCAmelCase_ : Any = [] # noqa: E741 for _ in range(20 ): l.append([int(A__ ) for x in f.readline().split()] ) UpperCAmelCase_ : Any = 0 # right for i in range(20 ): for j in range(17 ): UpperCAmelCase_ : Union[str, Any] = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: UpperCAmelCase_ : Any = temp # down for i in range(17 ): for j in range(20 ): UpperCAmelCase_ : List[Any] = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: UpperCAmelCase_ : Tuple = temp # diagonal 1 for i in range(17 ): for j in range(17 ): UpperCAmelCase_ : str = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: UpperCAmelCase_ : List[str] = temp # diagonal 2 for i in range(17 ): for j in range(3 ,20 ): UpperCAmelCase_ : List[Any] = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: UpperCAmelCase_ : List[str] = temp return maximum if __name__ == "__main__": print(solution())
268
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.g4dn.xlarge", "results": {"train_runtime": 6_5_0, "eval_accuracy": 0.6, "eval_loss": 0.9}, }, { "framework": "tensorflow", "script": "run_tf.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.g4dn.xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.3, "eval_loss": 0.9}, }, ] ) class UpperCamelCase ( unittest.TestCase ): def a_ ( self) -> Any: if self.framework == "pytorch": subprocess.run( f'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split(), encoding='utf-8', check=SCREAMING_SNAKE_CASE_, ) assert hasattr(self, 'env') def a_ ( self, lowerCAmelCase__=1) -> List[str]: # creates estimator return HuggingFace( entry_point=self.script, source_dir=self.env.test_path, role=self.env.role, image_uri=self.env.image_uri, base_job_name=f'{self.env.base_job_name}-single', instance_count=SCREAMING_SNAKE_CASE_, instance_type=self.instance_type, debugger_hook_config=SCREAMING_SNAKE_CASE_, hyperparameters={**self.env.hyperparameters, 'model_name_or_path': self.model_name_or_path}, metric_definitions=self.env.metric_definitions, py_version='py36', ) def a_ ( self, lowerCAmelCase__) -> str: TrainingJobAnalytics(SCREAMING_SNAKE_CASE_).export_csv(f'{self.env.test_path}/{job_name}_metrics.csv') def a_ ( self) -> Optional[int]: # create estimator snake_case_ = self.create_estimator() # run training estimator.fit() # result dataframe snake_case_ = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis snake_case_ = list(result_metrics_df[result_metrics_df.metric_name == 'eval_accuracy']['value']) snake_case_ = list(result_metrics_df[result_metrics_df.metric_name == 'eval_loss']['value']) # get train time from SageMaker job, this includes starting, preprocessing, stopping snake_case_ = ( Session().describe_training_job(estimator.latest_training_job.name).get('TrainingTimeInSeconds', 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['eval_accuracy'] for t in eval_accuracy) assert all(t <= self.results['eval_loss'] for t in eval_loss) # dump tests result into json file to share in PR with open(f'{estimator.latest_training_job.name}.json', 'w') as outfile: json.dump({'train_time': train_runtime, 'eval_accuracy': eval_accuracy, 'eval_loss': eval_loss}, SCREAMING_SNAKE_CASE_)
365
"""simple docstring""" from __future__ import annotations def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> list[str]: if partitions <= 0: raise ValueError('partitions must be a positive number!' ) if partitions > number_of_bytes: raise ValueError('partitions can not > number_of_bytes!' ) snake_case_ = number_of_bytes // partitions snake_case_ = [] for i in range(UpperCAmelCase ): snake_case_ = i * bytes_per_partition + 1 snake_case_ = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(f'{start_bytes}-{end_bytes}' ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
312
0
'''simple docstring''' from math import factorial class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: _lowerCAmelCase = real if isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowerCAmelCase = [1] * rank else: _lowerCAmelCase = rank def __repr__( self ) -> List[str]: return ( f'''{self.real}+''' f'''{'+'.join(str(_lowerCAmelCase )+'E'+str(n+1 )for n,dual in enumerate(self.duals ) )}''' ) def _snake_case ( self ) -> str: _lowerCAmelCase = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real , _lowerCAmelCase ) def __add__( self , _lowerCAmelCase ) -> Union[str, Any]: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return Dual(self.real + other , self.duals ) _lowerCAmelCase = self.duals.copy() _lowerCAmelCase = other.duals.copy() if len(_lowerCAmelCase ) > len(_lowerCAmelCase ): o_dual.extend([1] * (len(_lowerCAmelCase ) - len(_lowerCAmelCase )) ) elif len(_lowerCAmelCase ) < len(_lowerCAmelCase ): s_dual.extend([1] * (len(_lowerCAmelCase ) - len(_lowerCAmelCase )) ) _lowerCAmelCase = [] for i in range(len(_lowerCAmelCase ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real , _lowerCAmelCase ) __lowerCamelCase : str = __add__ def __sub__( self , _lowerCAmelCase ) -> Any: return self + other * -1 def __mul__( self , _lowerCAmelCase ) -> int: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowerCAmelCase = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other , _lowerCAmelCase ) _lowerCAmelCase = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real , _lowerCAmelCase ) __lowerCamelCase : str = __mul__ def __truediv__( self , _lowerCAmelCase ) -> List[Any]: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowerCAmelCase = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other , _lowerCAmelCase ) raise ValueError def __floordiv__( self , _lowerCAmelCase ) -> Tuple: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowerCAmelCase = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other , _lowerCAmelCase ) raise ValueError def __pow__( self , _lowerCAmelCase ) -> List[Any]: if n < 0 or isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise ValueError("power must be a positive integer" ) if n == 0: return 1 if n == 1: return self _lowerCAmelCase = self for _ in range(n - 1 ): x *= self return x def __a(SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Optional[int] ): '''simple docstring''' if not callable(SCREAMING_SNAKE_CASE_ ): raise ValueError("differentiate() requires a function as input for func" ) if not isinstance(SCREAMING_SNAKE_CASE_ , (float, int) ): raise ValueError("differentiate() requires a float as input for position" ) if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError("differentiate() requires an int as input for order" ) _lowerCAmelCase = Dual(SCREAMING_SNAKE_CASE_ , 1 ) _lowerCAmelCase = func(SCREAMING_SNAKE_CASE_ ) if order == 0: return result.real return result.duals[order - 1] * factorial(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod() def __a(SCREAMING_SNAKE_CASE_ : Any ): '''simple docstring''' return y**2 * y**4 print(differentiate(f, 9, 2))
158
'''simple docstring''' import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=14 , _lowerCAmelCase=7 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=False , _lowerCAmelCase=True , _lowerCAmelCase=99 , _lowerCAmelCase=32 , _lowerCAmelCase=4 , _lowerCAmelCase=4 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=512 , _lowerCAmelCase=0.02 , ) -> Any: _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = seq_length _lowerCAmelCase = is_training _lowerCAmelCase = use_input_mask _lowerCAmelCase = use_token_type_ids _lowerCAmelCase = use_labels _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = rotary_dim _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = initializer_range _lowerCAmelCase = None _lowerCAmelCase = vocab_size - 1 _lowerCAmelCase = vocab_size - 1 _lowerCAmelCase = vocab_size - 1 def _snake_case ( self ) -> str: _lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase = None if self.use_input_mask: _lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCAmelCase = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=_lowerCAmelCase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def _snake_case ( self ) -> Any: _lowerCAmelCase = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = config_and_inputs _lowerCAmelCase = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Dict: _lowerCAmelCase = 20 _lowerCAmelCase = model_class_name(_lowerCAmelCase ) _lowerCAmelCase = model.init_cache(input_ids.shape[0] , _lowerCAmelCase ) _lowerCAmelCase = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype="i4" ) _lowerCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) _lowerCAmelCase = model( input_ids[:, :-1] , attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase , position_ids=_lowerCAmelCase , ) _lowerCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) _lowerCAmelCase = model( input_ids[:, -1:] , attention_mask=_lowerCAmelCase , past_key_values=outputs_cache.past_key_values , position_ids=_lowerCAmelCase , ) _lowerCAmelCase = model(_lowerCAmelCase ) _lowerCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'''Max diff is {diff}''' ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> int: _lowerCAmelCase = 20 _lowerCAmelCase = model_class_name(_lowerCAmelCase ) _lowerCAmelCase = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) _lowerCAmelCase = model.init_cache(input_ids.shape[0] , _lowerCAmelCase ) _lowerCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) _lowerCAmelCase = model( input_ids[:, :-1] , attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase , position_ids=_lowerCAmelCase , ) _lowerCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) _lowerCAmelCase = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=_lowerCAmelCase , position_ids=_lowerCAmelCase , ) _lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) _lowerCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'''Max diff is {diff}''' ) @require_flax class lowerCAmelCase_ ( __magic_name__ ,__magic_name__ ,unittest.TestCase ): __lowerCamelCase : str = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () __lowerCamelCase : Optional[Any] = (FlaxGPTJForCausalLM,) if is_flax_available() else () def _snake_case ( self ) -> List[str]: _lowerCAmelCase = FlaxGPTJModelTester(self ) def _snake_case ( self ) -> List[str]: for model_class_name in self.all_model_classes: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _snake_case ( self ) -> int: for model_class_name in self.all_model_classes: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) @tooslow def _snake_case ( self ) -> Any: _lowerCAmelCase = GPTaTokenizer.from_pretrained("gpt2" , pad_token="<|endoftext|>" , padding_side="left" ) _lowerCAmelCase = tokenizer(["Hello this is a long string", "Hey"] , return_tensors="np" , padding=_lowerCAmelCase , truncation=_lowerCAmelCase ) _lowerCAmelCase = FlaxGPTJForCausalLM.from_pretrained("EleutherAI/gpt-j-6B" ) _lowerCAmelCase = False _lowerCAmelCase = model.config.eos_token_id _lowerCAmelCase = jax.jit(model.generate ) _lowerCAmelCase = jit_generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , pad_token_id=tokenizer.pad_token_id ).sequences _lowerCAmelCase = tokenizer.batch_decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) _lowerCAmelCase = [ "Hello this is a long string of text.\n\nI'm trying to get the text of the", "Hey, I'm a little late to the party. I'm going to", ] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) @is_pt_flax_cross_test def _snake_case ( self ) -> Optional[Any]: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs _lowerCAmelCase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) _lowerCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class _lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning _lowerCAmelCase = getattr(_lowerCAmelCase , _lowerCAmelCase ) _lowerCAmelCase , _lowerCAmelCase = pt_inputs["input_ids"].shape _lowerCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_lowerCAmelCase ): _lowerCAmelCase = 0 _lowerCAmelCase = 1 _lowerCAmelCase = 0 _lowerCAmelCase = 1 _lowerCAmelCase = pt_model_class(_lowerCAmelCase ).eval() _lowerCAmelCase = model_class(_lowerCAmelCase , dtype=jnp.floataa ) _lowerCAmelCase = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _lowerCAmelCase ) _lowerCAmelCase = fx_state with torch.no_grad(): _lowerCAmelCase = pt_model(**_lowerCAmelCase ).to_tuple() _lowerCAmelCase = fx_model(**_lowerCAmelCase ).to_tuple() self.assertEqual(len(_lowerCAmelCase ) , len(_lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_lowerCAmelCase , _lowerCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_lowerCAmelCase ) _lowerCAmelCase = model_class.from_pretrained(_lowerCAmelCase , from_pt=_lowerCAmelCase ) _lowerCAmelCase = fx_model_loaded(**_lowerCAmelCase ).to_tuple() self.assertEqual( len(_lowerCAmelCase ) , len(_lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(_lowerCAmelCase , _lowerCAmelCase ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def _snake_case ( self ) -> Dict: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs _lowerCAmelCase = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) _lowerCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class _lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning _lowerCAmelCase = getattr(_lowerCAmelCase , _lowerCAmelCase ) _lowerCAmelCase = pt_model_class(_lowerCAmelCase ).eval() _lowerCAmelCase = model_class(_lowerCAmelCase , dtype=jnp.floataa ) _lowerCAmelCase = load_flax_weights_in_pytorch_model(_lowerCAmelCase , fx_model.params ) _lowerCAmelCase , _lowerCAmelCase = pt_inputs["input_ids"].shape _lowerCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_lowerCAmelCase ): _lowerCAmelCase = 0 _lowerCAmelCase = 1 _lowerCAmelCase = 0 _lowerCAmelCase = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): _lowerCAmelCase = pt_model(**_lowerCAmelCase ).to_tuple() _lowerCAmelCase = fx_model(**_lowerCAmelCase ).to_tuple() self.assertEqual(len(_lowerCAmelCase ) , len(_lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_lowerCAmelCase , _lowerCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_lowerCAmelCase ) _lowerCAmelCase = pt_model_class.from_pretrained(_lowerCAmelCase , from_flax=_lowerCAmelCase ) with torch.no_grad(): _lowerCAmelCase = pt_model_loaded(**_lowerCAmelCase ).to_tuple() self.assertEqual( len(_lowerCAmelCase ) , len(_lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_lowerCAmelCase , _lowerCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def _snake_case ( self ) -> Union[str, Any]: for model_class_name in self.all_model_classes: _lowerCAmelCase = model_class_name.from_pretrained("EleutherAI/gpt-j-6B" ) _lowerCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(_lowerCAmelCase )
158
1
"""simple docstring""" import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class _SCREAMING_SNAKE_CASE( unittest.TestCase ): @require_torch def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = pipeline( task='''zero-shot-audio-classification''' ,model='''hf-internal-testing/tiny-clap-htsat-unfused''' ) __SCREAMING_SNAKE_CASE :Any = load_dataset('''ashraq/esc50''' ) __SCREAMING_SNAKE_CASE :int = dataset['''train''']['''audio'''][-1]['''array'''] __SCREAMING_SNAKE_CASE :Dict = audio_classifier(SCREAMING_SNAKE_CASE__ ,candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) ,[{'''score''': 0.5_0_1, '''label''': '''Sound of a dog'''}, {'''score''': 0.4_9_9, '''label''': '''Sound of vaccum cleaner'''}] ,) @unittest.skip('''No models are available in TF''' ) def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" pass @slow @require_torch def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = pipeline( task='''zero-shot-audio-classification''' ,model='''laion/clap-htsat-unfused''' ,) # This is an audio of a dog __SCREAMING_SNAKE_CASE :List[Any] = load_dataset('''ashraq/esc50''' ) __SCREAMING_SNAKE_CASE :Tuple = dataset['''train''']['''audio'''][-1]['''array'''] __SCREAMING_SNAKE_CASE :str = audio_classifier(SCREAMING_SNAKE_CASE__ ,candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) ,[ {'''score''': 0.9_9_9, '''label''': '''Sound of a dog'''}, {'''score''': 0.0_0_1, '''label''': '''Sound of vaccum cleaner'''}, ] ,) __SCREAMING_SNAKE_CASE :Dict = audio_classifier([audio] * 5 ,candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) ,[ [ {'''score''': 0.9_9_9, '''label''': '''Sound of a dog'''}, {'''score''': 0.0_0_1, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 ,) __SCREAMING_SNAKE_CASE :Union[str, Any] = audio_classifier( [audio] * 5 ,candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ,batch_size=5 ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) ,[ [ {'''score''': 0.9_9_9, '''label''': '''Sound of a dog'''}, {'''score''': 0.0_0_1, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 ,) @unittest.skip('''No models are available in TF''' ) def _UpperCamelCase ( self ) -> List[str]: """simple docstring""" pass
239
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "google/vivit-b-16x2-kinetics400": ( "https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json" ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : Optional[Any] = '''vivit''' def __init__( self ,SCREAMING_SNAKE_CASE__=2_24 ,SCREAMING_SNAKE_CASE__=32 ,SCREAMING_SNAKE_CASE__=[2, 16, 16] ,SCREAMING_SNAKE_CASE__=3 ,SCREAMING_SNAKE_CASE__=7_68 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=30_72 ,SCREAMING_SNAKE_CASE__="gelu_fast" ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=1E-06 ,SCREAMING_SNAKE_CASE__=True ,**SCREAMING_SNAKE_CASE__ ,) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_size __SCREAMING_SNAKE_CASE :List[str] = num_hidden_layers __SCREAMING_SNAKE_CASE :List[str] = num_attention_heads __SCREAMING_SNAKE_CASE :Any = intermediate_size __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_act __SCREAMING_SNAKE_CASE :int = hidden_dropout_prob __SCREAMING_SNAKE_CASE :Union[str, Any] = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :Dict = initializer_range __SCREAMING_SNAKE_CASE :Dict = layer_norm_eps __SCREAMING_SNAKE_CASE :Any = image_size __SCREAMING_SNAKE_CASE :Any = num_frames __SCREAMING_SNAKE_CASE :Any = tubelet_size __SCREAMING_SNAKE_CASE :Tuple = num_channels __SCREAMING_SNAKE_CASE :List[str] = qkv_bias super().__init__(**SCREAMING_SNAKE_CASE__ )
239
1
"""simple docstring""" from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging lowercase__ = logging.get_logger(__name__) def _snake_case ( lowercase__ , lowercase__ ): try: with open(lowercase__ , 'rb' ) as flax_state_f: _lowerCamelCase : Union[str, Any] = from_bytes(lowercase__ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowercase__ ) as f: if f.read().startswith('version' ): raise OSError( 'You seem to have cloned a repository without having git-lfs installed. Please' ' install git-lfs and run `git lfs install` followed by `git lfs pull` in the' ' folder you cloned.' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f'''Unable to convert {model_file} to Flax deserializable object. ''' ) return load_flax_weights_in_pytorch_model(lowercase__ , lowercase__ ) def _snake_case ( lowercase__ , lowercase__ ): try: import torch # noqa: F401 except ImportError: logger.error( 'Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights _lowerCamelCase : Union[str, Any] = flatten_dict(jax.tree_util.tree_map(lambda lowercase__ : x.dtype == jnp.bfloataa , lowercase__ ) ).values() if any(lowercase__ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) _lowerCamelCase : Any = jax.tree_util.tree_map( lambda lowercase__ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowercase__ ) _lowerCamelCase : Tuple = '' _lowerCamelCase : Union[str, Any] = flatten_dict(lowercase__ , sep='.' ) _lowerCamelCase : List[Any] = pt_model.state_dict() # keep track of unexpected & missing keys _lowerCamelCase : str = [] _lowerCamelCase : List[str] = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): _lowerCamelCase : int = flax_key_tuple.split('.' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: _lowerCamelCase : str = flax_key_tuple_array[:-1] + ['weight'] _lowerCamelCase : Optional[Any] = jnp.transpose(lowercase__ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": _lowerCamelCase : List[Any] = flax_key_tuple_array[:-1] + ['weight'] _lowerCamelCase : Tuple = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": _lowerCamelCase : str = flax_key_tuple_array[:-1] + ['weight'] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowercase__ ): _lowerCamelCase : List[str] = ( flax_key_tuple_string.replace('_0' , '.0' ) .replace('_1' , '.1' ) .replace('_2' , '.2' ) .replace('_3' , '.3' ) .replace('_4' , '.4' ) .replace('_5' , '.5' ) .replace('_6' , '.6' ) .replace('_7' , '.7' ) .replace('_8' , '.8' ) .replace('_9' , '.9' ) ) _lowerCamelCase : int = '.'.join(lowercase__ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f'''Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ''' f'''to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) else: # add weight to pytorch dict _lowerCamelCase : str = np.asarray(lowercase__ ) if not isinstance(lowercase__ , np.ndarray ) else flax_tensor _lowerCamelCase : List[str] = torch.from_numpy(lowercase__ ) # remove from missing keys missing_keys.remove(lowercase__ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowercase__ ) pt_model.load_state_dict(lowercase__ ) # re-transform missing_keys to list _lowerCamelCase : Dict = list(lowercase__ ) if len(lowercase__ ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' f''' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing''' f''' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture''' ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' f''' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect''' ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) if len(lowercase__ ) > 0: logger.warning( f'''Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly''' f''' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to''' ' use it for predictions and inference.' ) return pt_model
96
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class A__ ( A__ ): A__ = ( 'This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.' 'It takes two arguments named `image` which should be the original image, and `label` which should be a text ' 'describing the elements what should be identified in the segmentation mask. The tool returns the mask.' ) A__ = 'CIDAS/clipseg-rd64-refined' A__ = 'image_segmenter' A__ = CLIPSegForImageSegmentation A__ = ['image', 'text'] A__ = ['image'] def __init__( self : Any , *_a : Dict , **_a : str ) -> Any: '''simple docstring''' requires_backends(self , ['vision'] ) super().__init__(*_a , **_a ) def A ( self : int , _a : "Image" , _a : str ) -> Optional[Any]: '''simple docstring''' return self.pre_processor(text=[label] , images=[image] , padding=_a , return_tensors='pt' ) def A ( self : Dict , _a : Dict ) -> str: '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE =self.model(**_a ).logits return logits def A ( self : Any , _a : str ) -> Union[str, Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =outputs.cpu().detach().numpy() _SCREAMING_SNAKE_CASE =0 _SCREAMING_SNAKE_CASE =1 return Image.fromarray((array * 255).astype(np.uinta ) )
47
0
"""simple docstring""" import torch from transformers import AutoModel class _lowerCamelCase ( torch.nn.Module ): def __init__( self : List[Any] , UpperCamelCase : Tuple="sayef/fsner-bert-base-uncased" ) -> int: """simple docstring""" super(UpperCamelCase , self ).__init__() lowerCAmelCase__ : Any = AutoModel.from_pretrained(UpperCamelCase , return_dict=UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = torch.nn.CosineSimilarity(3 , 1E-0_8 ) lowerCAmelCase__ : Optional[Any] = torch.nn.Softmax(dim=1 ) def _lowerCAmelCase ( self : Any , **UpperCamelCase : Tuple ) -> Any: """simple docstring""" return self.bert(**UpperCamelCase ).last_hidden_state def _lowerCAmelCase ( self : int , UpperCamelCase : int ) -> List[str]: """simple docstring""" return token_embeddings.sum(2 , keepdim=UpperCamelCase ) def _lowerCAmelCase ( self : Any , UpperCamelCase : List[str] , UpperCamelCase : int , UpperCamelCase : Optional[Any]=1 ) -> Dict: """simple docstring""" return self.softmax(T * self.cos(UpperCamelCase , UpperCamelCase ) ) def _lowerCAmelCase ( self : Tuple , UpperCamelCase : Any , UpperCamelCase : Tuple ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = W_supports["""sizes"""].tolist() lowerCAmelCase__ : Any = W_supports["""start_token_id"""].item() lowerCAmelCase__ : List[str] = W_supports["""end_token_id"""].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] lowerCAmelCase__ : Union[str, Any] = self.BERT(**UpperCamelCase ) lowerCAmelCase__ : Any = self.BERT(**UpperCamelCase ) lowerCAmelCase__ : Optional[int] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[Any] = W_supports["""input_ids"""] == start_token_id lowerCAmelCase__ : List[Any] = W_supports["""input_ids"""] == end_token_id for i, size in enumerate(UpperCamelCase ): if i == 0: lowerCAmelCase__ : Optional[int] = 0 else: lowerCAmelCase__ : str = support_sizes[i - 1] lowerCAmelCase__ : Dict = S[s : s + size][start_token_masks[s : s + size]] lowerCAmelCase__ : List[Any] = S[s : s + size][end_token_masks[s : s + size]] lowerCAmelCase__ : Dict = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) lowerCAmelCase__ : Tuple = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: lowerCAmelCase__ : Optional[Any] = torch.vstack((p_starts, p_start) ) lowerCAmelCase__ : Union[str, Any] = torch.vstack((p_ends, p_end) ) else: lowerCAmelCase__ : int = p_start lowerCAmelCase__ : Optional[Any] = p_end return p_starts, p_ends
212
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder _A = """base_with_context""" def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> Dict: lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.FloatTensor(weights["""token_embedder"""]["""embedding"""] ) ) lowerCAmelCase__ : int = nn.Parameter( torch.FloatTensor(weights["""Embed_0"""]["""embedding"""] ) , requires_grad=__UpperCAmelCase ) for lyr_num, lyr in enumerate(model.encoders ): lowerCAmelCase__ : str = weights[f"""layers_{lyr_num}"""] lowerCAmelCase__ : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["""pre_attention_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : str = ly_weight["""attention"""] lowerCAmelCase__ : str = nn.Parameter(torch.FloatTensor(attention_weights["""query"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["""key"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["""value"""]["""kernel"""].T ) ) lowerCAmelCase__ : str = nn.Parameter(torch.FloatTensor(attention_weights["""out"""]["""kernel"""].T ) ) lowerCAmelCase__ : int = nn.Parameter(torch.FloatTensor(ly_weight["""pre_mlp_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wi_0"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wi_1"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wo"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[Any] = nn.Parameter(torch.FloatTensor(weights["""encoder_norm"""]["""scale"""] ) ) return model def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: lowerCAmelCase__ : Dict = nn.Parameter(torch.FloatTensor(weights["""input_proj"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[Any] = nn.Parameter( torch.FloatTensor(weights["""Embed_0"""]["""embedding"""] ) , requires_grad=__UpperCAmelCase ) for lyr_num, lyr in enumerate(model.encoders ): lowerCAmelCase__ : int = weights[f"""layers_{lyr_num}"""] lowerCAmelCase__ : Any = ly_weight["""attention"""] lowerCAmelCase__ : int = nn.Parameter(torch.FloatTensor(attention_weights["""query"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["""key"""]["""kernel"""].T ) ) lowerCAmelCase__ : str = nn.Parameter(torch.FloatTensor(attention_weights["""value"""]["""kernel"""].T ) ) lowerCAmelCase__ : Dict = nn.Parameter(torch.FloatTensor(attention_weights["""out"""]["""kernel"""].T ) ) lowerCAmelCase__ : Any = nn.Parameter( torch.FloatTensor(ly_weight["""pre_attention_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : Any = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wi_0"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wi_1"""]["""kernel"""].T ) ) lowerCAmelCase__ : Any = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wo"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["""pre_mlp_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.FloatTensor(weights["""encoder_norm"""]["""scale"""] ) ) return model def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> str: lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(weights["""time_emb_dense0"""]["""kernel"""].T ) ) lowerCAmelCase__ : Union[str, Any] = nn.Parameter(torch.FloatTensor(weights["""time_emb_dense1"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[str] = nn.Parameter( torch.FloatTensor(weights["""Embed_0"""]["""embedding"""] ) , requires_grad=__UpperCAmelCase ) lowerCAmelCase__ : Dict = nn.Parameter( torch.FloatTensor(weights["""continuous_inputs_projection"""]["""kernel"""].T ) ) for lyr_num, lyr in enumerate(model.decoders ): lowerCAmelCase__ : List[Any] = weights[f"""layers_{lyr_num}"""] lowerCAmelCase__ : Tuple = nn.Parameter( torch.FloatTensor(ly_weight["""pre_self_attention_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["""FiLMLayer_0"""]["""DenseGeneral_0"""]["""kernel"""].T ) ) lowerCAmelCase__ : Tuple = ly_weight["""self_attention"""] lowerCAmelCase__ : str = nn.Parameter(torch.FloatTensor(attention_weights["""query"""]["""kernel"""].T ) ) lowerCAmelCase__ : Dict = nn.Parameter(torch.FloatTensor(attention_weights["""key"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["""value"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["""out"""]["""kernel"""].T ) ) lowerCAmelCase__ : List[Any] = ly_weight["""MultiHeadDotProductAttention_0"""] lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["""query"""]["""kernel"""].T ) ) lowerCAmelCase__ : Dict = nn.Parameter(torch.FloatTensor(attention_weights["""key"""]["""kernel"""].T ) ) lowerCAmelCase__ : Any = nn.Parameter(torch.FloatTensor(attention_weights["""value"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["""out"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[int] = nn.Parameter( torch.FloatTensor(ly_weight["""pre_cross_attention_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["""pre_mlp_layer_norm"""]["""scale"""] ) ) lowerCAmelCase__ : Dict = nn.Parameter( torch.FloatTensor(ly_weight["""FiLMLayer_1"""]["""DenseGeneral_0"""]["""kernel"""].T ) ) lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wi_0"""]["""kernel"""].T ) ) lowerCAmelCase__ : int = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wi_1"""]["""kernel"""].T ) ) lowerCAmelCase__ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["""mlp"""]["""wo"""]["""kernel"""].T ) ) lowerCAmelCase__ : str = nn.Parameter(torch.FloatTensor(weights["""decoder_norm"""]["""scale"""] ) ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.FloatTensor(weights["""spec_out_dense"""]["""kernel"""].T ) ) return model def lowercase_ ( __UpperCAmelCase ) -> str: lowerCAmelCase__ : Optional[int] = checkpoints.load_tax_checkpoint(args.checkpoint_path ) lowerCAmelCase__ : Optional[int] = jnp.tree_util.tree_map(onp.array , __UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = [ """from __gin__ import dynamic_registration""", """from music_spectrogram_diffusion.models.diffusion import diffusion_utils""", """diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0""", """diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()""", ] lowerCAmelCase__ : Dict = os.path.join(args.checkpoint_path , """..""" , """config.gin""" ) lowerCAmelCase__ : Tuple = inference.parse_training_gin_file(__UpperCAmelCase , __UpperCAmelCase ) lowerCAmelCase__ : Any = inference.InferenceModel(args.checkpoint_path , __UpperCAmelCase ) lowerCAmelCase__ : List[Any] = DDPMScheduler(beta_schedule="""squaredcos_cap_v2""" , variance_type="""fixed_large""" ) lowerCAmelCase__ : List[Any] = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["""inputs"""] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="""gated-gelu""" , ) lowerCAmelCase__ : List[str] = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length["""targets_context"""] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="""gated-gelu""" , ) lowerCAmelCase__ : Optional[int] = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length["""targets_context"""] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) lowerCAmelCase__ : Optional[Any] = load_notes_encoder(ta_checkpoint["""target"""]["""token_encoder"""] , __UpperCAmelCase ) lowerCAmelCase__ : List[str] = load_continuous_encoder(ta_checkpoint["""target"""]["""continuous_encoder"""] , __UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = load_decoder(ta_checkpoint["""target"""]["""decoder"""] , __UpperCAmelCase ) lowerCAmelCase__ : Any = OnnxRuntimeModel.from_pretrained("""kashif/soundstream_mel_decoder""" ) lowerCAmelCase__ : Optional[Any] = SpectrogramDiffusionPipeline( notes_encoder=__UpperCAmelCase , continuous_encoder=__UpperCAmelCase , decoder=__UpperCAmelCase , scheduler=__UpperCAmelCase , melgan=__UpperCAmelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument("""--output_path""", default=None, type=str, required=True, help="""Path to the converted model.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument( """--checkpoint_path""", default=f"""{MODEL}/checkpoint_500000""", type=str, required=False, help="""Path to the original jax model checkpoint.""", ) _A = parser.parse_args() main(args)
212
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"ctrl": "https://huggingface.co/ctrl/resolve/main/config.json"} class UpperCAmelCase_ ( a__ ): lowercase__ = '''ctrl''' lowercase__ = ['''past_key_values'''] lowercase__ = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : List[str] , snake_case_ : Any=246_534 , snake_case_ : Optional[int]=256 , snake_case_ : Dict=1_280 , snake_case_ : Union[str, Any]=8_192 , snake_case_ : List[str]=48 , snake_case_ : Optional[Any]=16 , snake_case_ : Dict=0.1 , snake_case_ : Optional[int]=0.1 , snake_case_ : List[Any]=1e-6 , snake_case_ : Optional[int]=0.02 , snake_case_ : Tuple=True , **snake_case_ : Optional[Any] , ) -> Union[str, Any]: '''simple docstring''' A__ = vocab_size A__ = n_positions A__ = n_embd A__ = n_layer A__ = n_head A__ = dff A__ = resid_pdrop A__ = embd_pdrop A__ = layer_norm_epsilon A__ = initializer_range A__ = use_cache super().__init__(**_lowerCamelCase )
247
'''simple docstring''' import torch from torch import nn class _snake_case ( nn.Module ): def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=1 , _lowerCamelCase=False): super().__init__() UpperCAmelCase__ : List[Any] = n_token UpperCAmelCase__ : Tuple = d_embed UpperCAmelCase__ : str = d_proj UpperCAmelCase__ : str = cutoffs + [n_token] UpperCAmelCase__ : List[Any] = [0] + self.cutoffs UpperCAmelCase__ : Optional[Any] = div_val UpperCAmelCase__ : Optional[int] = self.cutoffs[0] UpperCAmelCase__ : Optional[int] = len(self.cutoffs) - 1 UpperCAmelCase__ : Union[str, Any] = self.shortlist_size + self.n_clusters if self.n_clusters > 0: UpperCAmelCase__ : int = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed)) UpperCAmelCase__ : Optional[Any] = nn.Parameter(torch.zeros(self.n_clusters)) UpperCAmelCase__ : int = nn.ModuleList() UpperCAmelCase__ : List[Any] = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs)): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(_lowerCamelCase , _lowerCamelCase))) else: self.out_projs.append(_lowerCamelCase) self.out_layers.append(nn.Linear(_lowerCamelCase , _lowerCamelCase)) else: for i in range(len(self.cutoffs)): UpperCAmelCase__ , UpperCAmelCase__ : List[str] = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase__ : Union[str, Any] = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(_lowerCamelCase , _lowerCamelCase))) self.out_layers.append(nn.Linear(_lowerCamelCase , r_idx - l_idx)) UpperCAmelCase__ : Optional[int] = keep_order def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase): if proj is None: UpperCAmelCase__ : Dict = nn.functional.linear(_lowerCamelCase , _lowerCamelCase , bias=_lowerCamelCase) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: UpperCAmelCase__ : Optional[int] = nn.functional.linear(_lowerCamelCase , proj.t().contiguous()) UpperCAmelCase__ : List[str] = nn.functional.linear(_lowerCamelCase , _lowerCamelCase , bias=_lowerCamelCase) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=False): if labels is not None: # Shift so that tokens < n predict n UpperCAmelCase__ : Optional[int] = hidden[..., :-1, :].contiguous() UpperCAmelCase__ : int = labels[..., 1:].contiguous() UpperCAmelCase__ : List[str] = hidden.view(-1 , hidden.size(-1)) UpperCAmelCase__ : Optional[int] = labels.view(-1) if hidden.size(0) != labels.size(0): raise RuntimeError("""Input and labels should have the same size in the batch dimension.""") else: UpperCAmelCase__ : Optional[int] = hidden.view(-1 , hidden.size(-1)) if self.n_clusters == 0: UpperCAmelCase__ : Tuple = self._compute_logit(_lowerCamelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0]) if labels is not None: UpperCAmelCase__ : Dict = labels != -100 UpperCAmelCase__ : Tuple = torch.zeros_like(_lowerCamelCase , dtype=hidden.dtype , device=hidden.device) UpperCAmelCase__ : List[Any] = ( -nn.functional.log_softmax(_lowerCamelCase , dim=-1)[mask].gather(1 , labels[mask].unsqueeze(1)).squeeze(1) ) else: UpperCAmelCase__ : List[str] = nn.functional.log_softmax(_lowerCamelCase , dim=-1) else: # construct weights and biases UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = [], [] for i in range(len(self.cutoffs)): if self.div_val == 1: UpperCAmelCase__ , UpperCAmelCase__ : int = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase__ : Dict = self.out_layers[0].weight[l_idx:r_idx] UpperCAmelCase__ : Any = self.out_layers[0].bias[l_idx:r_idx] else: UpperCAmelCase__ : Union[str, Any] = self.out_layers[i].weight UpperCAmelCase__ : Any = self.out_layers[i].bias if i == 0: UpperCAmelCase__ : Optional[Any] = torch.cat([weight_i, self.cluster_weight] , dim=0) UpperCAmelCase__ : List[Any] = torch.cat([bias_i, self.cluster_bias] , dim=0) weights.append(_lowerCamelCase) biases.append(_lowerCamelCase) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = weights[0], biases[0], self.out_projs[0] UpperCAmelCase__ : Optional[int] = self._compute_logit(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase) UpperCAmelCase__ : Union[str, Any] = nn.functional.log_softmax(_lowerCamelCase , dim=1) if labels is None: UpperCAmelCase__ : str = hidden.new_empty((head_logit.size(0), self.n_token)) else: UpperCAmelCase__ : Optional[Any] = torch.zeros_like(_lowerCamelCase , dtype=hidden.dtype , device=hidden.device) UpperCAmelCase__ : Optional[int] = 0 UpperCAmelCase__ : List[str] = [0] + self.cutoffs for i in range(len(_lowerCamelCase) - 1): UpperCAmelCase__ , UpperCAmelCase__ : Dict = cutoff_values[i], cutoff_values[i + 1] if labels is not None: UpperCAmelCase__ : List[str] = (labels >= l_idx) & (labels < r_idx) UpperCAmelCase__ : str = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue UpperCAmelCase__ : List[Any] = labels.index_select(0 , _lowerCamelCase) - l_idx UpperCAmelCase__ : List[str] = head_logprob.index_select(0 , _lowerCamelCase) UpperCAmelCase__ : Optional[Any] = hidden.index_select(0 , _lowerCamelCase) else: UpperCAmelCase__ : Any = hidden if i == 0: if labels is not None: UpperCAmelCase__ : List[Any] = head_logprob_i.gather(1 , target_i[:, None]).squeeze(1) else: UpperCAmelCase__ : Tuple = head_logprob[:, : self.cutoffs[0]] else: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : str = weights[i], biases[i], self.out_projs[i] UpperCAmelCase__ : int = self._compute_logit(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase) UpperCAmelCase__ : str = nn.functional.log_softmax(_lowerCamelCase , dim=1) UpperCAmelCase__ : int = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: UpperCAmelCase__ : Dict = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None]).squeeze(1) else: UpperCAmelCase__ : List[str] = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i UpperCAmelCase__ : Tuple = logprob_i if labels is not None: if (hasattr(self , """keep_order""") and self.keep_order) or keep_order: out.index_copy_(0 , _lowerCamelCase , -logprob_i) else: out[offset : offset + logprob_i.size(0)].copy_(-logprob_i) offset += logprob_i.size(0) return out def snake_case__ ( self , _lowerCamelCase): if self.n_clusters == 0: UpperCAmelCase__ : Union[str, Any] = self._compute_logit(_lowerCamelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0]) return nn.functional.log_softmax(_lowerCamelCase , dim=-1) else: # construct weights and biases UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = [], [] for i in range(len(self.cutoffs)): if self.div_val == 1: UpperCAmelCase__ , UpperCAmelCase__ : Any = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase__ : Union[str, Any] = self.out_layers[0].weight[l_idx:r_idx] UpperCAmelCase__ : Any = self.out_layers[0].bias[l_idx:r_idx] else: UpperCAmelCase__ : int = self.out_layers[i].weight UpperCAmelCase__ : List[str] = self.out_layers[i].bias if i == 0: UpperCAmelCase__ : List[Any] = torch.cat([weight_i, self.cluster_weight] , dim=0) UpperCAmelCase__ : Optional[int] = torch.cat([bias_i, self.cluster_bias] , dim=0) weights.append(_lowerCamelCase) biases.append(_lowerCamelCase) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : str = weights[0], biases[0], self.out_projs[0] UpperCAmelCase__ : List[Any] = self._compute_logit(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase) UpperCAmelCase__ : List[Any] = hidden.new_empty((head_logit.size(0), self.n_token)) UpperCAmelCase__ : int = nn.functional.log_softmax(_lowerCamelCase , dim=1) UpperCAmelCase__ : str = [0] + self.cutoffs for i in range(len(_lowerCamelCase) - 1): UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = cutoff_values[i], cutoff_values[i + 1] if i == 0: UpperCAmelCase__ : List[Any] = head_logprob[:, : self.cutoffs[0]] else: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = weights[i], biases[i], self.out_projs[i] UpperCAmelCase__ : Union[str, Any] = self._compute_logit(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase) UpperCAmelCase__ : List[str] = nn.functional.log_softmax(_lowerCamelCase , dim=1) UpperCAmelCase__ : Union[str, Any] = head_logprob[:, -i] + tail_logprob_i UpperCAmelCase__ : Dict = logprob_i return out
163
0
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> Optional[int]: __lowerCAmelCase = [] __lowerCAmelCase = set({"""(""", """[""", """{"""} ) __lowerCAmelCase = set({""")""", """]""", """}"""} ) __lowerCAmelCase = {"""{""": """}""", """[""": """]""", """(""": """)"""} for i in range(len(lowercase ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(lowercase ) == 0 or (len(lowercase ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(lowercase ) == 0 def _lowerCAmelCase ( ) -> Union[str, Any]: __lowerCAmelCase = input("""Enter sequence of brackets: """ ) if is_balanced(lowercase ): print(lowercase , """is balanced""" ) else: print(lowercase , """is not balanced""" ) if __name__ == "__main__": main()
370
'''simple docstring''' from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class _UpperCAmelCase ( lowerCAmelCase_ ): def lowerCamelCase__ ( self ): '''simple docstring''' return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def lowerCamelCase__ ( self ): '''simple docstring''' __lowerCAmelCase = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self ): '''simple docstring''' __lowerCAmelCase = self._create_example_records() __lowerCAmelCase = Dataset.from_list(__SCREAMING_SNAKE_CASE ) self.assertListEqual(dset.column_names,["""col_1""", """col_2"""] ) for i, r in enumerate(__SCREAMING_SNAKE_CASE ): self.assertDictEqual(__SCREAMING_SNAKE_CASE,example_records[i] ) def lowerCamelCase__ ( self ): '''simple docstring''' __lowerCAmelCase = self._create_example_records() __lowerCAmelCase = Dataset.from_list(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info,dset_from_dict.info ) def lowerCamelCase__ ( self ): # checks what happens with missing columns '''simple docstring''' __lowerCAmelCase = [{"""col_1""": 1}, {"""col_2""": """x"""}] __lowerCAmelCase = Dataset.from_list(__SCREAMING_SNAKE_CASE ) self.assertDictEqual(dset[0],{"""col_1""": 1} ) self.assertDictEqual(dset[1],{"""col_1""": None} ) # NB: first record is used for columns def lowerCamelCase__ ( self ): # checks if the type can be inferred from the second record '''simple docstring''' __lowerCAmelCase = [{"""col_1""": []}, {"""col_1""": [1, 2]}] __lowerCAmelCase = Dataset.from_list(__SCREAMING_SNAKE_CASE ) self.assertEqual(dset.info.features["""col_1"""],Sequence(Value("""int64""" ) ) ) def lowerCamelCase__ ( self ): '''simple docstring''' __lowerCAmelCase = Dataset.from_list([] ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ),0 ) self.assertListEqual(dset.column_names,[] )
46
0
import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class a__ ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Optional[Any] ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = [ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ ) ) def _lowercase ( self : Union[str, Any] ) ->List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = [ """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ ) ) def _lowercase ( self : str ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : Any = [ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", """unet/diffusion_pytorch_model.bin""", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(UpperCAmelCase__ ) ) def _lowercase ( self : Tuple ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = [ """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", ] self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ ) ) def _lowercase ( self : Tuple ) ->Any: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = [ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", # Removed: 'text_encoder/model.safetensors', """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertFalse(is_safetensors_compatible(UpperCAmelCase__ ) ) def _lowercase ( self : int ) ->Any: """simple docstring""" SCREAMING_SNAKE_CASE : str = [ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] SCREAMING_SNAKE_CASE : str = """fp16""" self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) ) def _lowercase ( self : List[str] ) ->Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = [ """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] SCREAMING_SNAKE_CASE : Optional[Any] = """fp16""" self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) ) def _lowercase ( self : Union[str, Any] ) ->Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = [ """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] SCREAMING_SNAKE_CASE : str = """fp16""" self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) ) def _lowercase ( self : Dict ) ->Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = [ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", """unet/diffusion_pytorch_model.fp16.bin""", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] SCREAMING_SNAKE_CASE : List[str] = """fp16""" self.assertFalse(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) ) def _lowercase ( self : str ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = [ """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", ] SCREAMING_SNAKE_CASE : Any = """fp16""" self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) ) def _lowercase ( self : str ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : str = [ """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", ] SCREAMING_SNAKE_CASE : Tuple = """fp16""" self.assertTrue(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) ) def _lowercase ( self : Optional[Any] ) ->Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = [ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", # 'text_encoder/model.fp16.safetensors', """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] SCREAMING_SNAKE_CASE : Optional[Any] = """fp16""" self.assertFalse(is_safetensors_compatible(UpperCAmelCase__ , variant=UpperCAmelCase__ ) )
245
import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ : """simple docstring""" def __init__( self : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any]=1_3 , UpperCAmelCase__ : List[str]=[3_0, 3_0] , UpperCAmelCase__ : Optional[int]=2 , UpperCAmelCase__ : Dict=3 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : int=3_2 , UpperCAmelCase__ : List[str]=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : int=3_7 , UpperCAmelCase__ : List[str]="gelu" , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : str=1_0 , UpperCAmelCase__ : Dict=0.02 , UpperCAmelCase__ : Any=3 , UpperCAmelCase__ : int=None , UpperCAmelCase__ : List[Any]=8 , UpperCAmelCase__ : Dict=1_0 , ) ->Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : str = parent SCREAMING_SNAKE_CASE : int = batch_size SCREAMING_SNAKE_CASE : str = image_size SCREAMING_SNAKE_CASE : List[Any] = patch_size SCREAMING_SNAKE_CASE : Any = num_channels SCREAMING_SNAKE_CASE : str = is_training SCREAMING_SNAKE_CASE : Dict = use_labels SCREAMING_SNAKE_CASE : List[Any] = hidden_size SCREAMING_SNAKE_CASE : Optional[int] = num_hidden_layers SCREAMING_SNAKE_CASE : Union[str, Any] = num_attention_heads SCREAMING_SNAKE_CASE : Dict = intermediate_size SCREAMING_SNAKE_CASE : int = hidden_act SCREAMING_SNAKE_CASE : str = hidden_dropout_prob SCREAMING_SNAKE_CASE : List[str] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : List[Any] = type_sequence_label_size SCREAMING_SNAKE_CASE : Optional[Any] = initializer_range SCREAMING_SNAKE_CASE : str = num_labels SCREAMING_SNAKE_CASE : Dict = scope SCREAMING_SNAKE_CASE : Optional[Any] = n_targets SCREAMING_SNAKE_CASE : Dict = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens SCREAMING_SNAKE_CASE : Tuple = (image_size[1] // patch_size) * (image_size[0] // patch_size) SCREAMING_SNAKE_CASE : int = num_patches + 1 + self.num_detection_tokens def _lowercase ( self : Tuple ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : str = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) SCREAMING_SNAKE_CASE : int = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) SCREAMING_SNAKE_CASE : str = [] for i in range(self.batch_size ): SCREAMING_SNAKE_CASE : List[Any] = {} SCREAMING_SNAKE_CASE : Any = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = torch.rand(self.n_targets , 4 , device=UpperCAmelCase__ ) labels.append(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = self.get_config() return config, pixel_values, labels def _lowercase ( self : Dict ) ->Optional[Any]: """simple docstring""" return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] ) ->Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = YolosModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() SCREAMING_SNAKE_CASE : Optional[Any] = model(UpperCAmelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def _lowercase ( self : str , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any] ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = YolosForObjectDetection(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() SCREAMING_SNAKE_CASE : List[Any] = model(pixel_values=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = model(UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) SCREAMING_SNAKE_CASE : int = model(pixel_values=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def _lowercase ( self : Dict ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[str] = config_and_inputs SCREAMING_SNAKE_CASE : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a__ ( UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : Tuple =(YolosModel, YolosForObjectDetection) if is_torch_available() else () UpperCAmelCase__ : Any =( {"""feature-extraction""": YolosModel, """object-detection""": YolosForObjectDetection} if is_torch_available() else {} ) UpperCAmelCase__ : Tuple =False UpperCAmelCase__ : int =False UpperCAmelCase__ : Tuple =False UpperCAmelCase__ : Optional[Any] =False def _lowercase ( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any=False ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = super()._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": SCREAMING_SNAKE_CASE : List[str] = [] for i in range(self.model_tester.batch_size ): SCREAMING_SNAKE_CASE : Tuple = {} SCREAMING_SNAKE_CASE : Union[str, Any] = torch.ones( size=(self.model_tester.n_targets,) , device=UpperCAmelCase__ , dtype=torch.long ) SCREAMING_SNAKE_CASE : str = torch.ones( self.model_tester.n_targets , 4 , device=UpperCAmelCase__ , dtype=torch.float ) labels.append(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : str = labels return inputs_dict def _lowercase ( self : Dict ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = YolosModelTester(self ) SCREAMING_SNAKE_CASE : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) ->List[str]: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : List[Any] ) ->int: """simple docstring""" pass def _lowercase ( self : Optional[int] ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : List[str] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def _lowercase ( self : List[Any] ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Tuple = model_class(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE : Tuple = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) ->Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) ->Dict: """simple docstring""" SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE : Optional[Any] = True # in YOLOS, the seq_len is different SCREAMING_SNAKE_CASE : Any = self.model_tester.expected_seq_len for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Optional[Any] = True SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = True SCREAMING_SNAKE_CASE : Optional[Any] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE : Union[str, Any] = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) SCREAMING_SNAKE_CASE : Union[str, Any] = outputs.attentions self.assertEqual(len(UpperCAmelCase__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] SCREAMING_SNAKE_CASE : Tuple = True SCREAMING_SNAKE_CASE : Tuple = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE : Dict = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) SCREAMING_SNAKE_CASE : Union[str, Any] = outputs.attentions self.assertEqual(len(UpperCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) SCREAMING_SNAKE_CASE : List[str] = len(UpperCAmelCase__ ) # Check attention is always last and order is fine SCREAMING_SNAKE_CASE : Optional[int] = True SCREAMING_SNAKE_CASE : Tuple = True SCREAMING_SNAKE_CASE : List[str] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE : List[Any] = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) SCREAMING_SNAKE_CASE : Optional[int] = 1 self.assertEqual(out_len + added_hidden_states , len(UpperCAmelCase__ ) ) SCREAMING_SNAKE_CASE : str = outputs.attentions self.assertEqual(len(UpperCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def _lowercase ( self : Any ) ->str: """simple docstring""" def check_hidden_states_output(UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE : List[Any] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE : List[str] = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) SCREAMING_SNAKE_CASE : Dict = outputs.hidden_states SCREAMING_SNAKE_CASE : str = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(UpperCAmelCase__ ) , UpperCAmelCase__ ) # YOLOS has a different seq_length SCREAMING_SNAKE_CASE : Tuple = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Any = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE : Any = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Any ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*UpperCAmelCase__ ) @slow def _lowercase ( self : str ) ->List[Any]: """simple docstring""" for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE : str = YolosModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def __lowercase ( ) -> List[Any]: SCREAMING_SNAKE_CASE : Tuple = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class a__ ( unittest.TestCase ): """simple docstring""" @cached_property def _lowercase ( self : int ) ->Union[str, Any]: """simple docstring""" return AutoImageProcessor.from_pretrained("""hustvl/yolos-small""" ) if is_vision_available() else None @slow def _lowercase ( self : List[Any] ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : int = YolosForObjectDetection.from_pretrained("""hustvl/yolos-small""" ).to(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Tuple = self.default_image_processor SCREAMING_SNAKE_CASE : Optional[Any] = prepare_img() SCREAMING_SNAKE_CASE : str = image_processor(images=UpperCAmelCase__ , return_tensors="""pt""" ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE : Optional[Any] = model(inputs.pixel_values ) # verify outputs SCREAMING_SNAKE_CASE : Union[str, Any] = torch.Size((1, 1_0_0, 9_2) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Any = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] , device=UpperCAmelCase__ , ) SCREAMING_SNAKE_CASE : Union[str, Any] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] , device=UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , UpperCAmelCase__ , atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , UpperCAmelCase__ , atol=1e-4 ) ) # verify postprocessing SCREAMING_SNAKE_CASE : int = image_processor.post_process_object_detection( UpperCAmelCase__ , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] SCREAMING_SNAKE_CASE : str = torch.tensor([0.99_94, 0.97_90, 0.99_64, 0.99_72, 0.98_61] ).to(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : str = [7_5, 7_5, 1_7, 6_3, 1_7] SCREAMING_SNAKE_CASE : List[str] = torch.tensor([3_35.06_09, 79.38_48, 3_75.42_16, 1_87.24_95] ).to(UpperCAmelCase__ ) self.assertEqual(len(results["""scores"""] ) , 5 ) self.assertTrue(torch.allclose(results["""scores"""] , UpperCAmelCase__ , atol=1e-4 ) ) self.assertSequenceEqual(results["""labels"""].tolist() , UpperCAmelCase__ ) self.assertTrue(torch.allclose(results["""boxes"""][0, :] , UpperCAmelCase__ ) )
245
1
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def _a ( self ) -> Dict: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights __UpperCamelCase =FlaxDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=A_ , cache_dir=A_ ) __UpperCamelCase =[t[-1] for t in os.walk(os.path.join(A_ , os.listdir(A_ )[0] , 'snapshots' ) )] __UpperCamelCase =[item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('.bin' ) for f in files ) @slow @require_flax class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def _a ( self ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=A_ ) __UpperCamelCase =( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =4 __UpperCamelCase =jax.device_count() __UpperCamelCase =num_samples * [prompt] __UpperCamelCase =pipeline.prepare_inputs(A_ ) # shard inputs and rng __UpperCamelCase =replicate(A_ ) __UpperCamelCase =jax.random.split(A_ , A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , A_ , jit=A_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.151_4745 ) < 1E-3 assert np.abs(np.abs(A_ , dtype=np.floataa ).sum() - 49947.875 ) < 5E-1 __UpperCamelCase =pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(A_ ) == num_samples def _a ( self ) -> List[str]: __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='flax' , safety_checker=A_ ) __UpperCamelCase =( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =50 __UpperCamelCase =jax.device_count() __UpperCamelCase =num_samples * [prompt] __UpperCamelCase =pipeline.prepare_inputs(A_ ) # shard inputs and rng __UpperCamelCase =replicate(A_ ) __UpperCamelCase =jax.random.split(A_ , A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , A_ , jit=A_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0565_2401) ) < 1E-3 assert np.abs((np.abs(A_ , dtype=np.floataa ).sum() - 2383808.2) ) < 5E-1 def _a ( self ) -> List[Any]: __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=A_ ) __UpperCamelCase =( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =50 __UpperCamelCase =jax.device_count() __UpperCamelCase =num_samples * [prompt] __UpperCamelCase =pipeline.prepare_inputs(A_ ) # shard inputs and rng __UpperCamelCase =replicate(A_ ) __UpperCamelCase =jax.random.split(A_ , A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , A_ , jit=A_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0400_3906) ) < 1E-3 assert np.abs((np.abs(A_ , dtype=np.floataa ).sum() - 2373516.75) ) < 5E-1 def _a ( self ) -> List[str]: __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa ) __UpperCamelCase =( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =50 __UpperCamelCase =jax.device_count() __UpperCamelCase =num_samples * [prompt] __UpperCamelCase =pipeline.prepare_inputs(A_ ) # shard inputs and rng __UpperCamelCase =replicate(A_ ) __UpperCamelCase =jax.random.split(A_ , A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , A_ , jit=A_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0400_3906) ) < 1E-3 assert np.abs((np.abs(A_ , dtype=np.floataa ).sum() - 2373516.75) ) < 5E-1 def _a ( self ) -> Tuple: __UpperCamelCase =FlaxDDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='scaled_linear' , set_alpha_to_one=A_ , steps_offset=1 , ) __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , scheduler=A_ , safety_checker=A_ , ) __UpperCamelCase =scheduler.create_state() __UpperCamelCase =scheduler_state __UpperCamelCase =( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =50 __UpperCamelCase =jax.device_count() __UpperCamelCase =num_samples * [prompt] __UpperCamelCase =pipeline.prepare_inputs(A_ ) # shard inputs and rng __UpperCamelCase =replicate(A_ ) __UpperCamelCase =jax.random.split(A_ , A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , A_ , jit=A_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_4504_3945) ) < 1E-3 assert np.abs((np.abs(A_ , dtype=np.floataa ).sum() - 2347693.5) ) < 5E-1 def _a ( self ) -> Any: __UpperCamelCase =( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) __UpperCamelCase =jax.device_count() __UpperCamelCase =num_samples * [prompt] __UpperCamelCase =jax.random.split(jax.random.PRNGKey(0 ) , A_ ) __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=A_ , ) __UpperCamelCase =replicate(A_ ) __UpperCamelCase =pipeline.prepare_inputs(A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , jit=A_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) __UpperCamelCase =images[2, 0, 256, 10:17, 1] # With memory efficient attention __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=A_ , use_memory_efficient_attention=A_ , ) __UpperCamelCase =replicate(A_ ) __UpperCamelCase =pipeline.prepare_inputs(A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipeline(A_ , A_ , A_ , jit=A_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) __UpperCamelCase =images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
365
from ....utils import logging _A = logging.get_logger(__name__) class UpperCAmelCase__ ( A_ ): """simple docstring""" def __init__( self , A_ , A_=None , A_=2048 ) -> Any: __UpperCamelCase =config.__dict__ __UpperCamelCase =modal_hidden_size if num_labels: __UpperCamelCase =num_labels
117
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : List[str] = logging.get_logger(__name__) __snake_case : str = { """microsoft/cvt-13""": """https://huggingface.co/microsoft/cvt-13/resolve/main/config.json""", # See all Cvt models at https://huggingface.co/models?filter=cvt } class A__(a_ ): """simple docstring""" _A : int = '''cvt''' def __init__( self , _lowercase=3 , _lowercase=[7, 3, 3] , _lowercase=[4, 2, 2] , _lowercase=[2, 1, 1] , _lowercase=[64, 192, 384] , _lowercase=[1, 3, 6] , _lowercase=[1, 2, 10] , _lowercase=[4.0, 4.0, 4.0] , _lowercase=[0.0, 0.0, 0.0] , _lowercase=[0.0, 0.0, 0.0] , _lowercase=[0.0, 0.0, 0.1] , _lowercase=[True, True, True] , _lowercase=[False, False, True] , _lowercase=["dw_bn", "dw_bn", "dw_bn"] , _lowercase=[3, 3, 3] , _lowercase=[1, 1, 1] , _lowercase=[2, 2, 2] , _lowercase=[1, 1, 1] , _lowercase=[1, 1, 1] , _lowercase=0.0_2 , _lowercase=1e-12 , **_lowercase , ) -> Union[str, Any]: super().__init__(**_lowercase ) a_ : Union[str, Any] = num_channels a_ : List[Any] = patch_sizes a_ : List[Any] = patch_stride a_ : List[Any] = patch_padding a_ : Optional[Any] = embed_dim a_ : List[str] = num_heads a_ : List[Any] = depth a_ : str = mlp_ratio a_ : List[Any] = attention_drop_rate a_ : Tuple = drop_rate a_ : Optional[int] = drop_path_rate a_ : str = qkv_bias a_ : List[Any] = cls_token a_ : Tuple = qkv_projection_method a_ : Union[str, Any] = kernel_qkv a_ : Optional[Any] = padding_kv a_ : Optional[Any] = stride_kv a_ : Tuple = padding_q a_ : Optional[Any] = stride_q a_ : List[Any] = initializer_range a_ : int = layer_norm_eps
248
import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format="""%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s""", datefmt="""%Y-%m-%d %H:%M:%S""", level=os.environ.get("""LOGLEVEL""", """INFO""").upper(), stream=sys.stdout, ) __snake_case : Any = logging.getLogger(__name__) __snake_case : Any = {"""facebook/bart-base""": BartForConditionalGeneration} __snake_case : Tuple = {"""facebook/bart-base""": BartTokenizer} def _UpperCAmelCase ( ): '''simple docstring''' a_ : List[str] = argparse.ArgumentParser(description="""Export Bart model + Beam Search to ONNX graph.""") parser.add_argument( """--validation_file""" , type=a__ , default=a__ , help="""A csv or a json file containing the validation data.""") parser.add_argument( """--max_length""" , type=a__ , default=5 , help="""The maximum total input sequence length after tokenization.""" , ) parser.add_argument( """--num_beams""" , type=a__ , default=a__ , help=( """Number of beams to use for evaluation. This argument will be """ """passed to ``model.generate``, which is used during ``evaluate`` and ``predict``.""" ) , ) parser.add_argument( """--model_name_or_path""" , type=a__ , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=a__ , ) parser.add_argument( """--config_name""" , type=a__ , default=a__ , help="""Pretrained config name or path if not the same as model_name""" , ) parser.add_argument( """--device""" , type=a__ , default="""cpu""" , help="""Device where the model will be run""" , ) parser.add_argument("""--output_file_path""" , type=a__ , default=a__ , help="""Where to store the final ONNX file.""") a_ : Any = parser.parse_args() return args def _UpperCAmelCase ( a__ , a__="cpu"): '''simple docstring''' a_ : Optional[int] = model_dict[model_name].from_pretrained(a__).to(a__) a_ : List[str] = tokenizer_dict[model_name].from_pretrained(a__) if model_name in ["facebook/bart-base"]: a_ : Tuple = 0 a_ : Optional[int] = None a_ : Union[str, Any] = 0 return huggingface_model, tokenizer def _UpperCAmelCase ( a__ , a__ , a__ , a__ , a__): '''simple docstring''' model.eval() a_ : Optional[Any] = None a_ : Optional[Any] = torch.jit.script(BARTBeamSearchGenerator(a__)) with torch.no_grad(): a_ : Any = """My friends are cool but they eat too many carbs.""" a_ : Dict = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=1_0_2_4 , return_tensors="""pt""").to(model.device) a_ : Optional[int] = model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , num_beams=a__ , max_length=a__ , early_stopping=a__ , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( a__ , ( inputs["""input_ids"""], inputs["""attention_mask"""], num_beams, max_length, model.config.decoder_start_token_id, ) , a__ , opset_version=1_4 , input_names=["""input_ids""", """attention_mask""", """num_beams""", """max_length""", """decoder_start_token_id"""] , output_names=["""output_ids"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """seq"""}, """output_ids""": {0: """batch""", 1: """seq_out"""}, } , example_outputs=a__ , ) logger.info("""Model exported to {}""".format(a__)) a_ : List[str] = remove_dup_initializers(os.path.abspath(a__)) logger.info("""Deduplicated and optimized model written to {}""".format(a__)) a_ : Union[str, Any] = onnxruntime.InferenceSession(a__) a_ : Any = ort_sess.run( a__ , { """input_ids""": inputs["""input_ids"""].cpu().numpy(), """attention_mask""": inputs["""attention_mask"""].cpu().numpy(), """num_beams""": np.array(a__), """max_length""": np.array(a__), """decoder_start_token_id""": np.array(model.config.decoder_start_token_id), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1e-3 , atol=1e-3) logger.info("""Model outputs from torch and ONNX Runtime are similar.""") logger.info("""Success.""") def _UpperCAmelCase ( ): '''simple docstring''' a_ : List[str] = parse_args() a_ : str = 5 a_ : Union[str, Any] = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.setLevel(logging.INFO) transformers.utils.logging.set_verbosity_error() a_ : int = torch.device(args.device) a_ , a_ : Optional[Any] = load_model_tokenizer(args.model_name_or_path , a__) if model.config.decoder_start_token_id is None: raise ValueError("""Make sure that `config.decoder_start_token_id` is correctly defined""") model.to(a__) if args.max_length: a_ : List[str] = args.max_length if args.num_beams: a_ : Optional[Any] = args.num_beams if args.output_file_path: a_ : Optional[int] = args.output_file_path else: a_ : Tuple = """BART.onnx""" logger.info("""Exporting model to ONNX""") export_and_validate_model(a__ , a__ , a__ , a__ , a__) if __name__ == "__main__": main()
248
1
import numpy as np import datasets _snake_case = ''' Compute the Mahalanobis Distance Mahalonobis distance is the distance between a point and a distribution. And not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance. It was introduced by Prof. P. C. Mahalanobis in 1936 and has been used in various statistical applications ever since [source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/] ''' _snake_case = '''\ @article{de2000mahalanobis, title={The mahalanobis distance}, author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L}, journal={Chemometrics and intelligent laboratory systems}, volume={50}, number={1}, pages={1--18}, year={2000}, publisher={Elsevier} } ''' _snake_case = ''' Args: X: List of datapoints to be compared with the `reference_distribution`. reference_distribution: List of datapoints from the reference distribution we want to compare to. Returns: mahalanobis: The Mahalonobis distance for each datapoint in `X`. Examples: >>> mahalanobis_metric = datasets.load_metric("mahalanobis") >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]]) >>> print(results) {\'mahalanobis\': array([0.5])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def _lowerCamelCase ( self: List[str] ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float" , id="sequence" ) , id="X" ), } ) , ) def _lowerCamelCase ( self: List[str] , __lowerCamelCase: int , __lowerCamelCase: Union[str, Any] ) -> List[str]: # convert to numpy arrays __UpperCAmelCase : int = np.array(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = np.array(__lowerCamelCase ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError("Expected `X` to be a 2D vector" ) if len(reference_distribution.shape ) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector" ) if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension" ) # Get mahalanobis distance for each prediction __UpperCAmelCase : str = X - np.mean(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = np.cov(reference_distribution.T ) try: __UpperCAmelCase : int = np.linalg.inv(__lowerCamelCase ) except np.linalg.LinAlgError: __UpperCAmelCase : Optional[int] = np.linalg.pinv(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = np.dot(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Optional[int] = np.dot(__lowerCamelCase , X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
342
import math _snake_case = 10 _snake_case = 7 _snake_case = BALLS_PER_COLOUR * NUM_COLOURS def _UpperCamelCase ( snake_case__ = 20 ) -> str: __UpperCAmelCase : Optional[Any] = math.comb(snake_case__, snake_case__ ) __UpperCAmelCase : List[Any] = math.comb(NUM_BALLS - BALLS_PER_COLOUR, snake_case__ ) __UpperCAmelCase : Dict = NUM_COLOURS * (1 - missing_colour / total) return f'''{result:.9f}''' if __name__ == "__main__": print(solution(20))
342
1
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } _UpperCamelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): for attribute in key.split('''.''' ): __lowerCAmelCase : List[str] = getattr(lowercase__ , lowercase__ ) if weight_type is not None: __lowerCAmelCase : Any = getattr(lowercase__ , lowercase__ ).shape else: __lowerCAmelCase : Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : str = value elif weight_type == "weight_g": __lowerCAmelCase : str = value elif weight_type == "weight_v": __lowerCAmelCase : List[Any] = value elif weight_type == "bias": __lowerCAmelCase : Tuple = value else: __lowerCAmelCase : Dict = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Dict = [] __lowerCAmelCase : int = fairseq_model.state_dict() __lowerCAmelCase : Any = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight __lowerCAmelCase : Any = None for name, value in fairseq_dict.items(): __lowerCAmelCase : Union[str, Any] = False if "conv_layers" in name: load_conv_layer( lowercase__ , lowercase__ , lowercase__ , lowercase__ , hf_model.config.feat_extract_norm == '''group''' , ) __lowerCAmelCase : Union[str, Any] = True elif name.split('''.''' )[0] == "proj": __lowerCAmelCase : List[Any] = fairseq_model.proj __lowerCAmelCase : Tuple = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCAmelCase : Optional[Any] = True if "*" in mapped_key: __lowerCAmelCase : str = name.split(lowercase__ )[0].split('''.''' )[-2] __lowerCAmelCase : Dict = mapped_key.replace('''*''' , lowercase__ ) if "weight_g" in name: __lowerCAmelCase : Tuple = '''weight_g''' elif "weight_v" in name: __lowerCAmelCase : List[Any] = '''weight_v''' elif "bias" in name: __lowerCAmelCase : int = '''bias''' elif "weight" in name: __lowerCAmelCase : Optional[int] = '''weight''' else: __lowerCAmelCase : str = None set_recursively(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) continue if not is_used: unused_weights.append(lowercase__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) return proj_weight def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Optional[Any] = full_name.split('''conv_layers.''' )[-1] __lowerCAmelCase : Optional[int] = name.split('''.''' ) __lowerCAmelCase : str = int(items[0] ) __lowerCAmelCase : int = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : str = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : str = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[int] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase__ ) def _lowercase ( lowercase__ ): __lowerCAmelCase, __lowerCAmelCase : Optional[Any] = emb.weight.shape __lowerCAmelCase : Dict = nn.Linear(lowercase__ , lowercase__ , bias=lowercase__ ) __lowerCAmelCase : Optional[Any] = emb.weight.data return lin_layer def _lowercase ( lowercase__ ): with open(lowercase__ , '''r''' , encoding='''utf-8''' ) as f: __lowerCAmelCase : Tuple = f.readlines() __lowerCAmelCase : Tuple = [line.split(''' ''' )[0] for line in lines] __lowerCAmelCase : Tuple = len(lowercase__ ) __lowerCAmelCase : List[str] = { '''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3, } vocab_dict.update(dict(zip(lowercase__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): __lowerCAmelCase : List[Any] = WavaVecaConfig.from_pretrained(lowercase__ ) __lowerCAmelCase : Union[str, Any] = SpeechaTextaConfig.from_pretrained( lowercase__ , vocab_size=lowercase__ , decoder_layers=lowercase__ , do_stable_layer_norm=lowercase__ ) __lowerCAmelCase : List[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowercase__ , return_attention_mask=lowercase__ , ) __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) __lowerCAmelCase : List[Any] = model[0].eval() # set weights for wav2vec2 encoder __lowerCAmelCase : List[Any] = WavaVecaModel(lowercase__ ) __lowerCAmelCase : Any = recursively_load_weights_wavaveca(model.encoder , lowercase__ ) __lowerCAmelCase : int = SpeechaTextaForCausalLM(lowercase__ ) __lowerCAmelCase, __lowerCAmelCase : int = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=lowercase__ ) # set output linear layer unexpected_keys.remove('''embed_out''' ) __lowerCAmelCase : Optional[Any] = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(f"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(f"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) __lowerCAmelCase : Optional[Any] = SpeechEncoderDecoderModel(encoder=lowercase__ , decoder=lowercase__ ) __lowerCAmelCase : List[str] = False # add projection layer __lowerCAmelCase : List[Any] = nn.Parameter(projection_layer.weight ) __lowerCAmelCase : Any = nn.Parameter(projection_layer.bias ) __lowerCAmelCase : Dict = create_vocab_dict(lowercase__ ) with open(os.path.join(lowercase__ , '''vocab.json''' ) , '''w''' ) as fp: json.dump(lowercase__ , lowercase__ ) __lowerCAmelCase : str = SpeechaTextaTokenizer(os.path.join(lowercase__ , '''vocab.json''' ) ) tokenizer.save_pretrained(lowercase__ ) __lowerCAmelCase : Any = hf_wavavec.config.to_dict() __lowerCAmelCase : List[Any] = tokenizer.pad_token_id __lowerCAmelCase : int = tokenizer.bos_token_id __lowerCAmelCase : str = tokenizer.eos_token_id __lowerCAmelCase : Optional[Any] = '''speech_to_text_2''' __lowerCAmelCase : Optional[int] = '''wav2vec2''' __lowerCAmelCase : int = SpeechEncoderDecoderConfig.from_dict(lowercase__ ) hf_wavavec.save_pretrained(lowercase__ ) feature_extractor.save_pretrained(lowercase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument( "--encoder_config_path", default="facebook/wav2vec2-large-lv60", type=str, help="Path to hf encoder wav2vec2 checkpoint config", ) parser.add_argument( "--decoder_config_path", default="facebook/s2t-small-mustc-en-fr-st", type=str, help="Path to hf decoder s2t checkpoint config", ) parser.add_argument("--vocab_size", default=1_0224, type=int, help="Vocab size of decoder") parser.add_argument("--num_decoder_layers", default=7, type=int, help="Number of decoder layers") _UpperCamelCase = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
275
from ....configuration_utils import PretrainedConfig from ....utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class __lowercase (_UpperCAmelCase ): _UpperCamelCase = """trajectory_transformer""" _UpperCamelCase = ["""past_key_values"""] _UpperCamelCase = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=100 , A_=5 , A_=1 , A_=1 , A_=249 , A_=6 , A_=17 , A_=25 , A_=4 , A_=4 , A_=128 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.0_006 , A_=512 , A_=0.02 , A_=1e-12 , A_=1 , A_=True , A_=1 , A_=5_0256 , A_=5_0256 , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase : Any = vocab_size __lowerCAmelCase : Tuple = action_weight __lowerCAmelCase : Tuple = reward_weight __lowerCAmelCase : Union[str, Any] = value_weight __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : str = block_size __lowerCAmelCase : Optional[Any] = action_dim __lowerCAmelCase : Union[str, Any] = observation_dim __lowerCAmelCase : Union[str, Any] = transition_dim __lowerCAmelCase : Dict = learning_rate __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Any = n_head __lowerCAmelCase : Optional[int] = n_embd __lowerCAmelCase : str = embd_pdrop __lowerCAmelCase : Dict = attn_pdrop __lowerCAmelCase : Optional[int] = resid_pdrop __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Optional[int] = layer_norm_eps __lowerCAmelCase : Any = kaiming_initializer_range __lowerCAmelCase : List[str] = use_cache super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ )
275
1
def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" def merge(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(SCREAMING_SNAKE_CASE ) <= 1: return collection lowercase__ = len(SCREAMING_SNAKE_CASE ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase = input('Enter numbers separated by a comma:\n').strip() lowerCAmelCase = [int(item) for item in user_input.split(',')] print(*merge_sort(unsorted), sep=',')
93
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = checkpoints.load_tax_checkpoint(SCREAMING_SNAKE_CASE ) lowercase__ = flatten_dict(SCREAMING_SNAKE_CASE ) return flax_params def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = {} lowercase__ = { '''token_embedder''': '''embeddings''', '''encoder_norm''': '''layernorm''', '''kernel''': '''weight''', '''.out''': '''.output''', '''scale''': '''weight''', '''embedders_0.pos_embedding''': '''row_embedder.weight''', '''embedders_1.pos_embedding''': '''column_embedder.weight''', } lowercase__ = { '''query''': '''attention.query''', '''key''': '''attention.key''', '''value''': '''attention.value''', '''output.dense''': '''output''', '''encoder_decoder_attention.o''': '''encoder_decoder_attention.attention.o''', '''pre_self_attention_layer_norm''': '''self_attention.layer_norm''', '''pre_cross_attention_layer_norm''': '''encoder_decoder_attention.layer_norm''', '''mlp.''': '''mlp.DenseReluDense.''', '''pre_mlp_layer_norm''': '''mlp.layer_norm''', '''self_attention.o''': '''self_attention.attention.o''', '''decoder.embeddings.embedding''': '''decoder.embed_tokens.weight''', '''decoder.relpos_bias.rel_embedding''': '''decoder.layer.0.self_attention.attention.relative_attention_bias.weight''', '''decoder.decoder_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.logits_dense.weight''': '''decoder.lm_head.weight''', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key lowercase__ = '''.'''.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): lowercase__ = new_key.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): lowercase__ = new_key.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number lowercase__ = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , SCREAMING_SNAKE_CASE ) lowercase__ = new_key.replace('''encoder''' , '''encoder.encoder''' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number lowercase__ = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , SCREAMING_SNAKE_CASE ) lowercase__ = flax_dict[key] lowercase__ = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): lowercase__ = torch.from_numpy(converted_dict[key].T ) else: lowercase__ = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=False ): """simple docstring""" lowercase__ = get_flax_param(SCREAMING_SNAKE_CASE ) if not use_large: lowercase__ = PixaStructVisionConfig() lowercase__ = PixaStructTextConfig() else: lowercase__ = PixaStructVisionConfig( hidden_size=15_36 , d_ff=39_68 , num_attention_heads=24 , num_hidden_layers=18 ) lowercase__ = PixaStructTextConfig(hidden_size=15_36 , d_ff=39_68 , num_heads=24 , num_layers=18 ) lowercase__ = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=SCREAMING_SNAKE_CASE ) lowercase__ = PixaStructForConditionalGeneration(SCREAMING_SNAKE_CASE ) lowercase__ = rename_and_convert_flax_params(SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) lowercase__ = AutoTokenizer.from_pretrained('''ybelkada/test-pix2struct-tokenizer''' ) lowercase__ = PixaStructImageProcessor() lowercase__ = PixaStructProcessor(image_processor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) if use_large: lowercase__ = 40_96 lowercase__ = True # mkdir if needed os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) model.save_pretrained(SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) print('''Model saved in {}'''.format(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--t5x_checkpoint_path', default=None, type=str, help='Path to the original T5x checkpoint.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--use_large', action='store_true', help='Use large model.') parser.add_argument('--is_vqa', action='store_true', help='Use large model.') lowerCAmelCase = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
93
1
'''simple docstring''' import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter lowercase_ = True except ImportError: lowercase_ = False lowercase_ = logging.get_logger(__name__) # pylint: disable=invalid-name def lowerCamelCase ( __lowerCamelCase : Namespace ) ->Optional[Any]: return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class a_ ( snake_case_ ): '''simple docstring''' @staticmethod def snake_case_( A ) -> Tuple: _SCREAMING_SNAKE_CASE = parser.add_parser("""add-new-model""" ) add_new_model_parser.add_argument("""--testing""" , action="""store_true""" , help="""If in testing mode.""" ) add_new_model_parser.add_argument("""--testing_file""" , type=A , help="""Configuration file on which to run.""" ) add_new_model_parser.add_argument( """--path""" , type=A , help="""Path to cookiecutter. Should only be used for testing purposes.""" ) add_new_model_parser.set_defaults(func=A ) def __init__( self , A , A , A=None , *A ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = testing _SCREAMING_SNAKE_CASE = testing_file _SCREAMING_SNAKE_CASE = path def snake_case_( self ) -> List[str]: warnings.warn( """The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. """ """It is not actively maintained anymore, so might give a result that won't pass all tests and quality """ """checks, you should use `transformers-cli add-new-model-like` instead.""" ) if not _has_cookiecutter: raise ImportError( """Model creation dependencies are required to use the `add_new_model` command. Install them by running """ """the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n""" ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory _SCREAMING_SNAKE_CASE = [directory for directory in os.listdir() if """cookiecutter-template-""" == directory[:22]] if len(A ) > 0: raise ValueError( """Several directories starting with `cookiecutter-template-` in current working directory. """ """Please clean your directory by removing all folders starting with `cookiecutter-template-` or """ """change your working directory.""" ) _SCREAMING_SNAKE_CASE = ( Path(A ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) _SCREAMING_SNAKE_CASE = path_to_transformer_root / """templates""" / """adding_a_new_model""" # Execute cookiecutter if not self._testing: cookiecutter(str(A ) ) else: with open(self._testing_file , """r""" ) as configuration_file: _SCREAMING_SNAKE_CASE = json.load(A ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=A , extra_context=A , ) _SCREAMING_SNAKE_CASE = [directory for directory in os.listdir() if """cookiecutter-template-""" in directory[:22]][0] # Retrieve configuration with open(directory + """/configuration.json""" , """r""" ) as configuration_file: _SCREAMING_SNAKE_CASE = json.load(A ) _SCREAMING_SNAKE_CASE = configuration["""lowercase_modelname"""] _SCREAMING_SNAKE_CASE = configuration["""generate_tensorflow_pytorch_and_flax"""] os.remove(f'{directory}/configuration.json' ) _SCREAMING_SNAKE_CASE = """PyTorch""" in generate_tensorflow_pytorch_and_flax _SCREAMING_SNAKE_CASE = """TensorFlow""" in generate_tensorflow_pytorch_and_flax _SCREAMING_SNAKE_CASE = """Flax""" in generate_tensorflow_pytorch_and_flax _SCREAMING_SNAKE_CASE = f'{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}' os.makedirs(A , exist_ok=A ) os.makedirs(f'{path_to_transformer_root}/tests/models/{lowercase_model_name}' , exist_ok=A ) # Tests require submodules as they have parent imports with open(f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py' , """w""" ): pass shutil.move( f'{directory}/__init__.py' , f'{model_dir}/__init__.py' , ) shutil.move( f'{directory}/configuration_{lowercase_model_name}.py' , f'{model_dir}/configuration_{lowercase_model_name}.py' , ) def remove_copy_lines(A ): with open(A , """r""" ) as f: _SCREAMING_SNAKE_CASE = f.readlines() with open(A , """w""" ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(A ) if output_pytorch: if not self._testing: remove_copy_lines(f'{directory}/modeling_{lowercase_model_name}.py' ) shutil.move( f'{directory}/modeling_{lowercase_model_name}.py' , f'{model_dir}/modeling_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/test_modeling_{lowercase_model_name}.py' , f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py' , ) else: os.remove(f'{directory}/modeling_{lowercase_model_name}.py' ) os.remove(f'{directory}/test_modeling_{lowercase_model_name}.py' ) if output_tensorflow: if not self._testing: remove_copy_lines(f'{directory}/modeling_tf_{lowercase_model_name}.py' ) shutil.move( f'{directory}/modeling_tf_{lowercase_model_name}.py' , f'{model_dir}/modeling_tf_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/test_modeling_tf_{lowercase_model_name}.py' , f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py' , ) else: os.remove(f'{directory}/modeling_tf_{lowercase_model_name}.py' ) os.remove(f'{directory}/test_modeling_tf_{lowercase_model_name}.py' ) if output_flax: if not self._testing: remove_copy_lines(f'{directory}/modeling_flax_{lowercase_model_name}.py' ) shutil.move( f'{directory}/modeling_flax_{lowercase_model_name}.py' , f'{model_dir}/modeling_flax_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/test_modeling_flax_{lowercase_model_name}.py' , f'{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py' , ) else: os.remove(f'{directory}/modeling_flax_{lowercase_model_name}.py' ) os.remove(f'{directory}/test_modeling_flax_{lowercase_model_name}.py' ) shutil.move( f'{directory}/{lowercase_model_name}.md' , f'{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md' , ) shutil.move( f'{directory}/tokenization_{lowercase_model_name}.py' , f'{model_dir}/tokenization_{lowercase_model_name}.py' , ) shutil.move( f'{directory}/tokenization_fast_{lowercase_model_name}.py' , f'{model_dir}/tokenization_{lowercase_model_name}_fast.py' , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(A , A , A ): # Create temp file _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = mkstemp() _SCREAMING_SNAKE_CASE = False with fdopen(A , """w""" ) as new_file: with open(A ) as old_file: for line in old_file: new_file.write(A ) if line_to_copy_below in line: _SCREAMING_SNAKE_CASE = True for line_to_copy in lines_to_copy: new_file.write(A ) if not line_found: raise ValueError(f'Line {line_to_copy_below} was not found in file.' ) # Copy the file permissions from the old file to the new file copymode(A , A ) # Remove original file remove(A ) # Move new file move(A , A ) def skip_units(A ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(A ): with open(A ) as datafile: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False for line in datafile: if "# To replace in: " in line and "##" not in line: _SCREAMING_SNAKE_CASE = line.split("""\"""" )[1] _SCREAMING_SNAKE_CASE = skip_units(A ) elif "# Below: " in line and "##" not in line: _SCREAMING_SNAKE_CASE = line.split("""\"""" )[1] _SCREAMING_SNAKE_CASE = skip_units(A ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(A , A , A ) _SCREAMING_SNAKE_CASE = [] elif "# Replace with" in line and "##" not in line: _SCREAMING_SNAKE_CASE = [] elif "##" not in line: lines_to_copy.append(A ) remove(A ) replace_in_files(f'{directory}/to_replace_{lowercase_model_name}.py' ) os.rmdir(A )
58
lowerCamelCase : Tuple = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def _SCREAMING_SNAKE_CASE ( lowercase : Optional[Any] , lowercase : int , lowercase : Optional[Any] , lowercase : List[Any] ): '''simple docstring''' lowerCamelCase_ = [False] * len(lowercase ) lowerCamelCase_ = [s] lowerCamelCase_ = True while queue: lowerCamelCase_ = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(lowercase ) lowerCamelCase_ = True lowerCamelCase_ = u return visited[t] def _SCREAMING_SNAKE_CASE ( lowercase : List[str] , lowercase : Tuple , lowercase : Tuple ): '''simple docstring''' lowerCamelCase_ = [-1] * (len(lowercase )) lowerCamelCase_ = 0 lowerCamelCase_ = [] lowerCamelCase_ = [i[:] for i in graph] # Record original cut, copy. while bfs(lowercase , lowercase , lowercase , lowercase ): lowerCamelCase_ = float('Inf' ) lowerCamelCase_ = sink while s != source: # Find the minimum value in select path lowerCamelCase_ = min(lowercase , graph[parent[s]][s] ) lowerCamelCase_ = parent[s] max_flow += path_flow lowerCamelCase_ = sink while v != source: lowerCamelCase_ = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow lowerCamelCase_ = parent[v] for i in range(len(lowercase ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
204
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase__ = { """configuration_git""": ["""GIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GitConfig""", """GitVisionConfig"""], """processing_git""": ["""GitProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ """GIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GitForCausalLM""", """GitModel""", """GitPreTrainedModel""", """GitVisionModel""", ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
30
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase__ = { """configuration_git""": ["""GIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GitConfig""", """GitVisionConfig"""], """processing_git""": ["""GitProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ """GIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GitForCausalLM""", """GitModel""", """GitPreTrainedModel""", """GitVisionModel""", ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
30
1
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : List[str] = [ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' UpperCAmelCase : List[Any] = [ """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Dict = [ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", """unet/diffusion_pytorch_model.bin""", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Dict = [ """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", ] self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Optional[int] = [ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", # Removed: 'text_encoder/model.safetensors', """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertFalse(is_safetensors_compatible(_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Dict = [ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] UpperCAmelCase : int = """fp16""" self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' UpperCAmelCase : Union[str, Any] = [ """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] UpperCAmelCase : Optional[Any] = """fp16""" self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : Dict = [ """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] UpperCAmelCase : Any = """fp16""" self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : Optional[Any] = [ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", """unet/diffusion_pytorch_model.fp16.bin""", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] UpperCAmelCase : Any = """fp16""" self.assertFalse(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' UpperCAmelCase : str = [ """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", ] UpperCAmelCase : str = """fp16""" self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase : Tuple = [ """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", ] UpperCAmelCase : Dict = """fp16""" self.assertTrue(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Any = [ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", # 'text_encoder/model.fp16.safetensors', """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] UpperCAmelCase : Optional[int] = """fp16""" self.assertFalse(is_safetensors_compatible(_SCREAMING_SNAKE_CASE , variant=_SCREAMING_SNAKE_CASE ) )
109
"""simple docstring""" A: int = range(2, 2_0 + 1) A: Any = [1_0**k for k in range(ks[-1] + 1)] A: dict[int, dict[int, list[list[int]]]] = {} def _snake_case ( UpperCamelCase : Dict , UpperCamelCase : Any , UpperCamelCase : Any , UpperCamelCase : int ): UpperCAmelCase : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) UpperCAmelCase : str = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) UpperCAmelCase , UpperCAmelCase : str = 0, 0 UpperCAmelCase : Optional[Any] = n - i UpperCAmelCase : Optional[int] = memo.get(UpperCamelCase ) if sub_memo is not None: UpperCAmelCase : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase : Tuple = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase : int = _k break if max_jump >= 0: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[Any] = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase : List[str] = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): UpperCAmelCase , UpperCAmelCase : Union[str, Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: UpperCAmelCase : int = [] else: UpperCAmelCase : List[str] = {c: []} UpperCAmelCase : str = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase , UpperCAmelCase : List[str] = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase , UpperCAmelCase : int = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase : Dict = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase : str = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def _snake_case ( UpperCamelCase : int , UpperCamelCase : Tuple , UpperCamelCase : Union[str, Any] , UpperCamelCase : Any ): if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase : List[str] = i UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase : Optional[int] = ds_c + ds_b diff += addend UpperCAmelCase : str = 0 for j in range(UpperCamelCase ): UpperCAmelCase : Any = a_i[j] + addend UpperCAmelCase , UpperCAmelCase : Any = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def _snake_case ( UpperCamelCase : Union[str, Any] , UpperCamelCase : str , UpperCamelCase : Optional[int] ): for j in range(UpperCamelCase , len(UpperCamelCase ) ): UpperCAmelCase : Optional[int] = digits[j] + addend if s >= 10: UpperCAmelCase , UpperCAmelCase : int = divmod(UpperCamelCase , 10 ) UpperCAmelCase : str = addend // 10 + quotient else: UpperCAmelCase : Any = s UpperCAmelCase : Union[str, Any] = addend // 10 if addend == 0: break while addend > 0: UpperCAmelCase , UpperCAmelCase : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def _snake_case ( UpperCamelCase : int = 10**15 ): UpperCAmelCase : Dict = [1] UpperCAmelCase : int = 1 UpperCAmelCase : Tuple = 0 while True: UpperCAmelCase , UpperCAmelCase : Tuple = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase : Any = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(f"""{solution() = }""")
109
1
def __UpperCAmelCase ( __a : list[int] ,__a : int ) -> bool: """simple docstring""" _a : List[str] = len(__a ) _a : Optional[int] = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _a : Optional[Any] = True # sum is not zero and set is empty then false for i in range(1 ,required_sum + 1 ): _a : Dict = False for i in range(1 ,arr_len + 1 ): for j in range(1 ,required_sum + 1 ): if arr[i - 1] > j: _a : Union[str, Any] = subset[i - 1][j] if arr[i - 1] <= j: _a : Tuple = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
15
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def __UpperCAmelCase ( ) -> Optional[Any]: """simple docstring""" _a : int = ArgumentParser('''Accelerate CLI tool''' ,usage='''accelerate <command> [<args>]''' ,allow_abbrev=__a ) _a : Optional[int] = parser.add_subparsers(help='''accelerate command helpers''' ) # Register commands get_config_parser(subparsers=__a ) env_command_parser(subparsers=__a ) launch_command_parser(subparsers=__a ) tpu_command_parser(subparsers=__a ) test_command_parser(subparsers=__a ) # Let's go _a : Dict = parser.parse_args() if not hasattr(__a ,'''func''' ): parser.print_help() exit(1 ) # Run args.func(__a ) if __name__ == "__main__": main()
15
1
'''simple docstring''' from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging __UpperCAmelCase =logging.get_logger(__name__) __UpperCAmelCase ={ "google/umt5-small": "https://huggingface.co/google/umt5-small/resolve/main/config.json", # See all umt5 models at https://huggingface.co/models?filter=umt5 } class a__ ( UpperCAmelCase__ ): lowerCamelCase : Any ="umt5" lowerCamelCase : Dict =["past_key_values"] def __init__( self : Optional[Any] , a : Optional[Any]=25_01_12 , a : str=5_12 , a : Union[str, Any]=64 , a : Union[str, Any]=10_24 , a : Dict=8 , a : Any=None , a : str=6 , a : Optional[int]=32 , a : List[str]=1_28 , a : Optional[int]=0.1 , a : Any=1e-6 , a : List[Any]=1.0 , a : int="gated-gelu" , a : str=True , a : Dict=True , a : Optional[int]="T5Tokenizer" , a : List[Any]=True , a : Optional[Any]=0 , a : Any=1 , a : Union[str, Any]=0 , **a : str , ): """simple docstring""" super().__init__( is_encoder_decoder=a , tokenizer_class=a , tie_word_embeddings=a , pad_token_id=a , eos_token_id=a , decoder_start_token_id=a , **a , ) __lowerCamelCase = vocab_size __lowerCamelCase = d_model __lowerCamelCase = d_kv __lowerCamelCase = d_ff __lowerCamelCase = num_layers __lowerCamelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __lowerCamelCase = num_heads __lowerCamelCase = relative_attention_num_buckets __lowerCamelCase = relative_attention_max_distance __lowerCamelCase = dropout_rate __lowerCamelCase = layer_norm_epsilon __lowerCamelCase = initializer_factor __lowerCamelCase = feed_forward_proj __lowerCamelCase = use_cache __lowerCamelCase = self.feed_forward_proj.split('''-''' ) __lowerCamelCase = act_info[-1] __lowerCamelCase = act_info[0] == '''gated''' if len(a ) > 1 and act_info[0] != "gated" or len(a ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) if feed_forward_proj == "gated-gelu": __lowerCamelCase = '''gelu_new''' @property def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" return self.d_model @property def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" return self.num_heads @property def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return self.num_layers class a__ ( UpperCAmelCase__ ): @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.inputs def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" __lowerCamelCase = { '''input_ids''': {0: '''batch''', 1: '''encoder_sequence'''}, '''attention_mask''': {0: '''batch''', 1: '''encoder_sequence'''}, } if self.use_past: __lowerCamelCase = '''past_encoder_sequence + sequence''' __lowerCamelCase = {0: '''batch'''} __lowerCamelCase = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: __lowerCamelCase = {0: '''batch''', 1: '''decoder_sequence'''} __lowerCamelCase = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(a , direction='''inputs''' ) return common_inputs @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.default_onnx_opset def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return 13 @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return 5e-4
67
"""simple docstring""" def a__ ( _SCREAMING_SNAKE_CASE = 4_000_000 ): """simple docstring""" UpperCamelCase = [] UpperCamelCase , UpperCamelCase = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(_SCREAMING_SNAKE_CASE ) UpperCamelCase , UpperCamelCase = b, a + b return sum(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(f'''{solution() = }''')
153
0
from __future__ import annotations __lowerCamelCase : Tuple = [True] * 100_0001 __lowerCamelCase : Tuple = 2 while i * i <= 100_0000: if seive[i]: for j in range(i * i, 100_0001, i): __lowerCamelCase : List[str] = False i += 1 def SCREAMING_SNAKE_CASE ( snake_case_ : int ): return seive[n] def SCREAMING_SNAKE_CASE ( snake_case_ : int ): return any(digit in "02468" for digit in str(snake_case_ ) ) def SCREAMING_SNAKE_CASE ( snake_case_ : int = 1000000 ): snake_case__ : int = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(snake_case_ ) and not contains_an_even_digit(snake_case_ ): snake_case__ : List[str] = str(snake_case_ ) snake_case__ : int = [int(str_num[j:] + str_num[:j] ) for j in range(len(snake_case_ ) )] if all(is_prime(snake_case_ ) for i in list_nums ): result.append(snake_case_ ) return result def SCREAMING_SNAKE_CASE ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"{len(find_circular_primes()) = }")
366
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self : Union[str, Any] , __A : str , __A : Optional[Any]=1_3 , __A : Dict=7 , __A : List[str]=True , __A : Any=True , __A : str=True , __A : Optional[Any]=True , __A : List[str]=9_9 , __A : Dict=3_2 , __A : Tuple=2 , __A : Tuple=4 , __A : Dict=3_7 , __A : Tuple="gelu" , __A : Any=0.1 , __A : str=0.1 , __A : int=5_1_2 , __A : Union[str, Any]=1_6 , __A : Optional[int]=2 , __A : Union[str, Any]=0.0_2 , __A : Tuple=3 , __A : Union[str, Any]=4 , __A : Optional[int]=None , ): snake_case__ : Optional[int] = parent snake_case__ : Optional[Any] = 1_3 snake_case__ : int = 7 snake_case__ : Optional[int] = True snake_case__ : Optional[Any] = True snake_case__ : List[str] = True snake_case__ : int = True snake_case__ : Optional[int] = 9_9 snake_case__ : Union[str, Any] = 3_8_4 snake_case__ : Optional[Any] = 2 snake_case__ : Union[str, Any] = 4 snake_case__ : Any = 3_7 snake_case__ : Any = "gelu" snake_case__ : str = 0.1 snake_case__ : Optional[Any] = 0.1 snake_case__ : Union[str, Any] = 5_1_2 snake_case__ : Optional[Any] = 1_6 snake_case__ : List[Any] = 2 snake_case__ : Optional[int] = 0.0_2 snake_case__ : Dict = 3 snake_case__ : Any = 4 snake_case__ : int = 1_2_8 snake_case__ : Dict = 2 snake_case__ : Any = 9 snake_case__ : List[str] = 1 snake_case__ : List[Any] = None def _lowercase ( self : List[str] ): snake_case__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ : str = None if self.use_input_mask: snake_case__ : str = random_attention_mask([self.batch_size, self.seq_length] ) snake_case__ : Union[str, Any] = None if self.use_token_type_ids: snake_case__ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case__ : Optional[Any] = None snake_case__ : Any = None snake_case__ : Tuple = None if self.use_labels: snake_case__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case__ : int = ids_tensor([self.batch_size] , self.num_choices ) snake_case__ : int = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__A , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self : Dict , __A : Dict , __A : Dict , __A : Union[str, Any] , __A : Optional[int] , __A : Any , __A : Union[str, Any] , __A : Tuple ): snake_case__ : Optional[int] = TFConvBertModel(config=__A ) snake_case__ : Optional[int] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} snake_case__ : List[str] = [input_ids, input_mask] snake_case__ : Union[str, Any] = model(__A ) snake_case__ : str = model(__A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Union[str, Any] , __A : List[Any] , __A : Any , __A : Union[str, Any] , __A : int , __A : Optional[Any] , __A : Dict , __A : Optional[int] ): snake_case__ : List[str] = TFConvBertForMaskedLM(config=__A ) snake_case__ : Any = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } snake_case__ : int = model(__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : Tuple , __A : Union[str, Any] , __A : List[Any] , __A : Any , __A : List[Any] , __A : List[Any] , __A : Optional[int] , __A : List[str] ): snake_case__ : Any = self.num_labels snake_case__ : List[Any] = TFConvBertForSequenceClassification(config=__A ) snake_case__ : Any = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } snake_case__ : Optional[int] = model(__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : int , __A : List[Any] , __A : Union[str, Any] , __A : Optional[Any] , __A : List[Any] , __A : Union[str, Any] , __A : Union[str, Any] , __A : Optional[int] ): snake_case__ : Optional[Any] = self.num_choices snake_case__ : Any = TFConvBertForMultipleChoice(config=__A ) snake_case__ : Optional[int] = tf.tile(tf.expand_dims(__A , 1 ) , (1, self.num_choices, 1) ) snake_case__ : Optional[Any] = tf.tile(tf.expand_dims(__A , 1 ) , (1, self.num_choices, 1) ) snake_case__ : Optional[int] = tf.tile(tf.expand_dims(__A , 1 ) , (1, self.num_choices, 1) ) snake_case__ : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } snake_case__ : Optional[Any] = model(__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowercase ( self : List[str] , __A : Tuple , __A : str , __A : Union[str, Any] , __A : Union[str, Any] , __A : Any , __A : int , __A : Tuple ): snake_case__ : Dict = self.num_labels snake_case__ : str = TFConvBertForTokenClassification(config=__A ) snake_case__ : Optional[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } snake_case__ : List[str] = model(__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : Optional[int] , __A : Union[str, Any] , __A : List[Any] , __A : List[str] , __A : Any , __A : Any , __A : Optional[int] , __A : Optional[Any] ): snake_case__ : Any = TFConvBertForQuestionAnswering(config=__A ) snake_case__ : List[str] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } snake_case__ : int = model(__A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self : Any ): snake_case__ : List[Any] = self.prepare_config_and_inputs() ( ( snake_case__ ), ( snake_case__ ), ( snake_case__ ), ( snake_case__ ), ( snake_case__ ), ( snake_case__ ), ( snake_case__ ), ) : List[str] = config_and_inputs snake_case__ : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): """simple docstring""" a_ = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) a_ = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) a_ = False a_ = False a_ = False def _lowercase ( self : int ): snake_case__ : Optional[Any] = TFConvBertModelTester(self ) snake_case__ : List[str] = ConfigTester(self , config_class=__A , hidden_size=3_7 ) def _lowercase ( self : List[Any] ): self.config_tester.run_common_tests() def _lowercase ( self : Any ): snake_case__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def _lowercase ( self : Union[str, Any] ): snake_case__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__A ) def _lowercase ( self : Dict ): snake_case__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__A ) def _lowercase ( self : Optional[Any] ): snake_case__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__A ) def _lowercase ( self : Optional[int] ): snake_case__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__A ) def _lowercase ( self : Dict ): snake_case__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__A ) @slow def _lowercase ( self : Dict ): snake_case__, snake_case__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : int = True snake_case__ : int = True if hasattr(__A , "use_cache" ): snake_case__ : Optional[Any] = True snake_case__ : Dict = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) snake_case__ : List[str] = getattr(self.model_tester , "key_length" , __A ) for model_class in self.all_model_classes: snake_case__ : Tuple = self._prepare_for_class(__A , __A ) snake_case__ : List[str] = model_class(__A ) snake_case__ : List[Any] = len(model(__A ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__A , saved_model=__A ) snake_case__ : str = os.path.join(__A , "saved_model" , "1" ) snake_case__ : str = tf.keras.models.load_model(__A ) snake_case__ : Optional[Any] = model(__A ) if self.is_encoder_decoder: snake_case__ : Tuple = outputs["encoder_hidden_states"] snake_case__ : str = outputs["encoder_attentions"] else: snake_case__ : Dict = outputs["hidden_states"] snake_case__ : Tuple = outputs["attentions"] self.assertEqual(len(__A ) , __A ) snake_case__ : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__A ) , __A ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def _lowercase ( self : Tuple ): snake_case__ : Optional[Any] = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(__A ) def _lowercase ( self : List[str] ): snake_case__, snake_case__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : Optional[Any] = True snake_case__ : List[Any] = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) snake_case__ : int = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) snake_case__ : Any = getattr(self.model_tester , "key_length" , __A ) snake_case__ : List[Any] = getattr(self.model_tester , "key_length" , __A ) def check_decoder_attentions_output(__A : Optional[int] ): snake_case__ : Optional[Any] = len(__A ) self.assertEqual(out_len % 2 , 0 ) snake_case__ : Optional[int] = outputs.decoder_attentions self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(__A : Any ): snake_case__ : List[Any] = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: snake_case__ : Optional[int] = True snake_case__ : Any = False snake_case__ : Dict = model_class(__A ) snake_case__ : List[Any] = model(self._prepare_for_class(__A , __A ) ) snake_case__ : Dict = len(__A ) self.assertEqual(config.output_hidden_states , __A ) check_encoder_attentions_output(__A ) if self.is_encoder_decoder: snake_case__ : str = model_class(__A ) snake_case__ : List[Any] = model(self._prepare_for_class(__A , __A ) ) self.assertEqual(config.output_hidden_states , __A ) check_decoder_attentions_output(__A ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] snake_case__ : Optional[int] = True snake_case__ : Optional[Any] = model_class(__A ) snake_case__ : Union[str, Any] = model(self._prepare_for_class(__A , __A ) ) self.assertEqual(config.output_hidden_states , __A ) check_encoder_attentions_output(__A ) # Check attention is always last and order is fine snake_case__ : Optional[int] = True snake_case__ : List[Any] = True snake_case__ : Any = model_class(__A ) snake_case__ : str = model(self._prepare_for_class(__A , __A ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__A ) ) self.assertEqual(model.config.output_hidden_states , __A ) check_encoder_attentions_output(__A ) @require_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self : int ): snake_case__ : int = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) snake_case__ : int = tf.constant([[0, 1, 2, 3, 4, 5]] ) snake_case__ : str = model(__A )[0] snake_case__ : int = [1, 6, 7_6_8] self.assertEqual(output.shape , __A ) snake_case__ : List[Any] = tf.constant( [ [ [-0.0_3_4_7_5_4_9_3, -0.4_6_8_6_0_3_4, -0.3_0_6_3_8_8_3_2], [0.2_2_6_3_7_2_4_8, -0.2_6_9_8_8_6_4_6, -0.7_4_2_3_4_2_4], [0.1_0_3_2_4_8_6_8, -0.4_5_0_1_3_5_0_8, -0.5_8_2_8_0_7_8_4], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __A , atol=1e-4 )
286
0
import argparse import os import re import packaging.version A__ : Dict = '''examples/''' A__ : Any = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } A__ : Any = { '''init''': '''src/transformers/__init__.py''', '''setup''': '''setup.py''', } A__ : Any = '''README.md''' def UpperCamelCase( __UpperCamelCase : int ,__UpperCamelCase : List[Any] ,__UpperCamelCase : List[Any] ): with open(__UpperCamelCase ,'''r''' ,encoding='''utf-8''' ,newline='''\n''' ) as f: lowerCAmelCase_ : Tuple = f.read() lowerCAmelCase_ , lowerCAmelCase_ : Dict = REPLACE_PATTERNS[pattern] lowerCAmelCase_ : Tuple = replace.replace('''VERSION''' ,__UpperCamelCase ) lowerCAmelCase_ : Optional[int] = re_pattern.sub(__UpperCamelCase ,__UpperCamelCase ) with open(__UpperCamelCase ,'''w''' ,encoding='''utf-8''' ,newline='''\n''' ) as f: f.write(__UpperCamelCase ) def UpperCamelCase( __UpperCamelCase : Union[str, Any] ): for folder, directories, fnames in os.walk(__UpperCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(__UpperCamelCase ,__UpperCamelCase ) ,__UpperCamelCase ,pattern='''examples''' ) def UpperCamelCase( __UpperCamelCase : int ,__UpperCamelCase : List[Any]=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) if not patch: update_version_in_examples(__UpperCamelCase ) def UpperCamelCase( ): lowerCAmelCase_ : List[str] = '''🤗 Transformers currently provides the following architectures''' lowerCAmelCase_ : List[Any] = '''1. Want to contribute a new model?''' with open(__UpperCamelCase ,'''r''' ,encoding='''utf-8''' ,newline='''\n''' ) as f: lowerCAmelCase_ : Union[str, Any] = f.readlines() # Find the start of the list. lowerCAmelCase_ : int = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowerCAmelCase_ : str = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowerCAmelCase_ : int = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' ,'''https://huggingface.co/docs/transformers/model_doc''' ,) index += 1 with open(__UpperCamelCase ,'''w''' ,encoding='''utf-8''' ,newline='''\n''' ) as f: f.writelines(__UpperCamelCase ) def UpperCamelCase( ): with open(REPLACE_FILES['''init'''] ,'''r''' ) as f: lowerCAmelCase_ : Optional[Any] = f.read() lowerCAmelCase_ : Dict = REPLACE_PATTERNS['''init'''][0].search(__UpperCamelCase ).groups()[0] return packaging.version.parse(__UpperCamelCase ) def UpperCamelCase( __UpperCamelCase : Dict=False ): lowerCAmelCase_ : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowerCAmelCase_ : List[str] = default_version.base_version elif patch: lowerCAmelCase_ : int = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowerCAmelCase_ : int = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowerCAmelCase_ : Optional[Any] = input(f"""Which version are you releasing? [{default_version}]""" ) if len(__UpperCamelCase ) == 0: lowerCAmelCase_ : List[str] = default_version print(f"""Updating version to {version}.""" ) global_version_update(__UpperCamelCase ,patch=__UpperCamelCase ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def UpperCamelCase( ): lowerCAmelCase_ : Any = get_version() lowerCAmelCase_ : int = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowerCAmelCase_ : Optional[Any] = current_version.base_version # Check with the user we got that right. lowerCAmelCase_ : Optional[Any] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(__UpperCamelCase ) == 0: lowerCAmelCase_ : int = dev_version print(f"""Updating version to {version}.""" ) global_version_update(__UpperCamelCase ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": A__ : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') A__ : Optional[int] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
103
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _a ( snake_case_ , snake_case_ ): """simple docstring""" @register_to_config def __init__( self : Dict , UpperCAmelCase : int = 768 , ): super().__init__() A_ = nn.Parameter(torch.zeros(1 , UpperCAmelCase ) ) A_ = nn.Parameter(torch.ones(1 , UpperCAmelCase ) ) def __A ( self : str , UpperCAmelCase : Optional[Union[str, torch.device]] = None , UpperCAmelCase : Optional[torch.dtype] = None , ): A_ = nn.Parameter(self.mean.to(UpperCAmelCase ).to(UpperCAmelCase ) ) A_ = nn.Parameter(self.std.to(UpperCAmelCase ).to(UpperCAmelCase ) ) return self def __A ( self : Dict , UpperCAmelCase : List[Any] ): A_ = (embeds - self.mean) * 1.0 / self.std return embeds def __A ( self : int , UpperCAmelCase : int ): A_ = (embeds * self.std) + self.mean return embeds
312
0
'''simple docstring''' def UpperCamelCase_( snake_case : int = 5_0 ): '''simple docstring''' snake_case_ = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F"{solution() = }")
363
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Optional[int] = { "edbeeching/decision-transformer-gym-hopper-medium": ( "https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json" ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class _snake_case ( lowercase_ ): lowerCAmelCase_ : Dict = "decision_transformer" lowerCAmelCase_ : List[Any] = ["past_key_values"] lowerCAmelCase_ : Tuple = { "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , a__=17 , a__=4 , a__=128 , a__=4_096 , a__=True , a__=1 , a__=1_024 , a__=3 , a__=1 , a__=None , a__="relu" , a__=0.1 , a__=0.1 , a__=0.1 , a__=1e-5 , a__=0.0_2 , a__=True , a__=True , a__=50_256 , a__=50_256 , a__=False , a__=False , **a__ , ) -> Optional[int]: '''simple docstring''' snake_case_ = state_dim snake_case_ = act_dim snake_case_ = hidden_size snake_case_ = max_ep_len snake_case_ = action_tanh snake_case_ = vocab_size snake_case_ = n_positions snake_case_ = n_layer snake_case_ = n_head snake_case_ = n_inner snake_case_ = activation_function snake_case_ = resid_pdrop snake_case_ = embd_pdrop snake_case_ = attn_pdrop snake_case_ = layer_norm_epsilon snake_case_ = initializer_range snake_case_ = scale_attn_weights snake_case_ = use_cache snake_case_ = scale_attn_by_inverse_layer_idx snake_case_ = reorder_and_upcast_attn snake_case_ = bos_token_id snake_case_ = eos_token_id super().__init__(bos_token_id=a__ , eos_token_id=a__ , **a__ )
92
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowercase : List[Any] = { "configuration_mobilebert": [ "MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MobileBertConfig", "MobileBertOnnxConfig", ], "tokenization_mobilebert": ["MobileBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : str = ["MobileBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : List[str] = [ "MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "MobileBertForMaskedLM", "MobileBertForMultipleChoice", "MobileBertForNextSentencePrediction", "MobileBertForPreTraining", "MobileBertForQuestionAnswering", "MobileBertForSequenceClassification", "MobileBertForTokenClassification", "MobileBertLayer", "MobileBertModel", "MobileBertPreTrainedModel", "load_tf_weights_in_mobilebert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[int] = [ "TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFMobileBertForMaskedLM", "TFMobileBertForMultipleChoice", "TFMobileBertForNextSentencePrediction", "TFMobileBertForPreTraining", "TFMobileBertForQuestionAnswering", "TFMobileBertForSequenceClassification", "TFMobileBertForTokenClassification", "TFMobileBertMainLayer", "TFMobileBertModel", "TFMobileBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys _lowercase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
239
'''simple docstring''' def lowerCamelCase ( UpperCAmelCase__ : list ) -> list: if any(not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) or x < 0 for x in sequence ): raise TypeError("""Sequence must be list of non-negative integers""" ) for _ in range(len(UpperCAmelCase__ ) ): for i, (rod_upper, rod_lower) in enumerate(zip(UpperCAmelCase__ , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
239
1
'''simple docstring''' from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES _lowerCamelCase : List[str] = logging.get_logger(__name__) _lowerCamelCase : Any = OrderedDict( [ # Base model mapping ('albert', 'FlaxAlbertModel'), ('bart', 'FlaxBartModel'), ('beit', 'FlaxBeitModel'), ('bert', 'FlaxBertModel'), ('big_bird', 'FlaxBigBirdModel'), ('blenderbot', 'FlaxBlenderbotModel'), ('blenderbot-small', 'FlaxBlenderbotSmallModel'), ('clip', 'FlaxCLIPModel'), ('distilbert', 'FlaxDistilBertModel'), ('electra', 'FlaxElectraModel'), ('gpt-sw3', 'FlaxGPT2Model'), ('gpt2', 'FlaxGPT2Model'), ('gpt_neo', 'FlaxGPTNeoModel'), ('gptj', 'FlaxGPTJModel'), ('longt5', 'FlaxLongT5Model'), ('marian', 'FlaxMarianModel'), ('mbart', 'FlaxMBartModel'), ('mt5', 'FlaxMT5Model'), ('opt', 'FlaxOPTModel'), ('pegasus', 'FlaxPegasusModel'), ('regnet', 'FlaxRegNetModel'), ('resnet', 'FlaxResNetModel'), ('roberta', 'FlaxRobertaModel'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormModel'), ('roformer', 'FlaxRoFormerModel'), ('t5', 'FlaxT5Model'), ('vision-text-dual-encoder', 'FlaxVisionTextDualEncoderModel'), ('vit', 'FlaxViTModel'), ('wav2vec2', 'FlaxWav2Vec2Model'), ('whisper', 'FlaxWhisperModel'), ('xglm', 'FlaxXGLMModel'), ('xlm-roberta', 'FlaxXLMRobertaModel'), ] ) _lowerCamelCase : Optional[Any] = OrderedDict( [ # Model for pre-training mapping ('albert', 'FlaxAlbertForPreTraining'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForPreTraining'), ('big_bird', 'FlaxBigBirdForPreTraining'), ('electra', 'FlaxElectraForPreTraining'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('t5', 'FlaxT5ForConditionalGeneration'), ('wav2vec2', 'FlaxWav2Vec2ForPreTraining'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) _lowerCamelCase : int = OrderedDict( [ # Model for Masked LM mapping ('albert', 'FlaxAlbertForMaskedLM'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForMaskedLM'), ('big_bird', 'FlaxBigBirdForMaskedLM'), ('distilbert', 'FlaxDistilBertForMaskedLM'), ('electra', 'FlaxElectraForMaskedLM'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) _lowerCamelCase : List[str] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('bart', 'FlaxBartForConditionalGeneration'), ('blenderbot', 'FlaxBlenderbotForConditionalGeneration'), ('blenderbot-small', 'FlaxBlenderbotSmallForConditionalGeneration'), ('encoder-decoder', 'FlaxEncoderDecoderModel'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('marian', 'FlaxMarianMTModel'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('pegasus', 'FlaxPegasusForConditionalGeneration'), ('t5', 'FlaxT5ForConditionalGeneration'), ] ) _lowerCamelCase : int = OrderedDict( [ # Model for Image-classsification ('beit', 'FlaxBeitForImageClassification'), ('regnet', 'FlaxRegNetForImageClassification'), ('resnet', 'FlaxResNetForImageClassification'), ('vit', 'FlaxViTForImageClassification'), ] ) _lowerCamelCase : int = OrderedDict( [ ('vision-encoder-decoder', 'FlaxVisionEncoderDecoderModel'), ] ) _lowerCamelCase : Optional[int] = OrderedDict( [ # Model for Causal LM mapping ('bart', 'FlaxBartForCausalLM'), ('bert', 'FlaxBertForCausalLM'), ('big_bird', 'FlaxBigBirdForCausalLM'), ('electra', 'FlaxElectraForCausalLM'), ('gpt-sw3', 'FlaxGPT2LMHeadModel'), ('gpt2', 'FlaxGPT2LMHeadModel'), ('gpt_neo', 'FlaxGPTNeoForCausalLM'), ('gptj', 'FlaxGPTJForCausalLM'), ('opt', 'FlaxOPTForCausalLM'), ('roberta', 'FlaxRobertaForCausalLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForCausalLM'), ('xglm', 'FlaxXGLMForCausalLM'), ('xlm-roberta', 'FlaxXLMRobertaForCausalLM'), ] ) _lowerCamelCase : Optional[Any] = OrderedDict( [ # Model for Sequence Classification mapping ('albert', 'FlaxAlbertForSequenceClassification'), ('bart', 'FlaxBartForSequenceClassification'), ('bert', 'FlaxBertForSequenceClassification'), ('big_bird', 'FlaxBigBirdForSequenceClassification'), ('distilbert', 'FlaxDistilBertForSequenceClassification'), ('electra', 'FlaxElectraForSequenceClassification'), ('mbart', 'FlaxMBartForSequenceClassification'), ('roberta', 'FlaxRobertaForSequenceClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForSequenceClassification'), ('roformer', 'FlaxRoFormerForSequenceClassification'), ('xlm-roberta', 'FlaxXLMRobertaForSequenceClassification'), ] ) _lowerCamelCase : Any = OrderedDict( [ # Model for Question Answering mapping ('albert', 'FlaxAlbertForQuestionAnswering'), ('bart', 'FlaxBartForQuestionAnswering'), ('bert', 'FlaxBertForQuestionAnswering'), ('big_bird', 'FlaxBigBirdForQuestionAnswering'), ('distilbert', 'FlaxDistilBertForQuestionAnswering'), ('electra', 'FlaxElectraForQuestionAnswering'), ('mbart', 'FlaxMBartForQuestionAnswering'), ('roberta', 'FlaxRobertaForQuestionAnswering'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForQuestionAnswering'), ('roformer', 'FlaxRoFormerForQuestionAnswering'), ('xlm-roberta', 'FlaxXLMRobertaForQuestionAnswering'), ] ) _lowerCamelCase : Union[str, Any] = OrderedDict( [ # Model for Token Classification mapping ('albert', 'FlaxAlbertForTokenClassification'), ('bert', 'FlaxBertForTokenClassification'), ('big_bird', 'FlaxBigBirdForTokenClassification'), ('distilbert', 'FlaxDistilBertForTokenClassification'), ('electra', 'FlaxElectraForTokenClassification'), ('roberta', 'FlaxRobertaForTokenClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForTokenClassification'), ('roformer', 'FlaxRoFormerForTokenClassification'), ('xlm-roberta', 'FlaxXLMRobertaForTokenClassification'), ] ) _lowerCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ('albert', 'FlaxAlbertForMultipleChoice'), ('bert', 'FlaxBertForMultipleChoice'), ('big_bird', 'FlaxBigBirdForMultipleChoice'), ('distilbert', 'FlaxDistilBertForMultipleChoice'), ('electra', 'FlaxElectraForMultipleChoice'), ('roberta', 'FlaxRobertaForMultipleChoice'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMultipleChoice'), ('roformer', 'FlaxRoFormerForMultipleChoice'), ('xlm-roberta', 'FlaxXLMRobertaForMultipleChoice'), ] ) _lowerCamelCase : int = OrderedDict( [ ('bert', 'FlaxBertForNextSentencePrediction'), ] ) _lowerCamelCase : Union[str, Any] = OrderedDict( [ ('speech-encoder-decoder', 'FlaxSpeechEncoderDecoderModel'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ] ) _lowerCamelCase : Any = OrderedDict( [ ('whisper', 'FlaxWhisperForAudioClassification'), ] ) _lowerCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) _lowerCamelCase : Dict = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) _lowerCamelCase : Tuple = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) _lowerCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) _lowerCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) _lowerCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) _lowerCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) _lowerCamelCase : int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) _lowerCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) _lowerCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) _lowerCamelCase : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) _lowerCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) _lowerCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) _lowerCamelCase : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_MAPPING _lowerCamelCase : Optional[Any] = auto_class_update(FlaxAutoModel) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_PRETRAINING_MAPPING _lowerCamelCase : List[str] = auto_class_update(FlaxAutoModelForPreTraining, head_doc='pretraining') class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING _lowerCamelCase : List[Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc='causal language modeling') class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_MASKED_LM_MAPPING _lowerCamelCase : List[str] = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='masked language modeling') class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING _lowerCamelCase : Tuple = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='sequence-to-sequence language modeling', checkpoint_for_example='t5-base' ) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _lowerCamelCase : Tuple = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='sequence classification' ) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING _lowerCamelCase : Any = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='question answering') class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING _lowerCamelCase : str = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='token classification' ) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING _lowerCamelCase : Tuple = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='multiple choice') class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING _lowerCamelCase : List[Any] = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='next sentence prediction' ) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _lowerCamelCase : Union[str, Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc='image classification' ) class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING _lowerCamelCase : Optional[int] = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='vision-to-text modeling') class __UpperCAmelCase ( _BaseAutoModelClass ): '''simple docstring''' __lowerCAmelCase = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING _lowerCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='sequence-to-sequence speech-to-text modeling' )
357
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging _lowerCamelCase : List[Any] = logging.get_logger(__name__) def __a ( UpperCAmelCase ) ->List[int]: """simple docstring""" if isinstance(UpperCAmelCase , np.ndarray ): return list(tensor.shape ) A = tf.shape(UpperCAmelCase ) if tensor.shape == tf.TensorShape(UpperCAmelCase ): return dynamic A = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(UpperCAmelCase )] def __a ( UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None ) ->tf.Tensor: """simple docstring""" return tf.nn.softmax(logits=logits + 1E-9 , axis=UpperCAmelCase , name=UpperCAmelCase ) def __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=1E-5 , UpperCAmelCase=-1 ) ->str: """simple docstring""" if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(UpperCAmelCase , UpperCAmelCase ): raise NotImplementedError("""Only 1D weight and bias tensors are supported for now, with only a single axis.""" ) # Get mean and variance on the axis to be normalized A , A = tf.nn.moments(UpperCAmelCase , axes=[axis] , keepdims=UpperCAmelCase ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis A = [1] * inputs.shape.rank A = shape_list(UpperCAmelCase )[axis] A = tf.reshape(UpperCAmelCase , UpperCAmelCase ) A = tf.reshape(UpperCAmelCase , UpperCAmelCase ) # Compute layer normalization using the batch_normalization # function. A = tf.nn.batch_normalization( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , offset=UpperCAmelCase , scale=UpperCAmelCase , variance_epsilon=UpperCAmelCase , ) return outputs def __a ( UpperCAmelCase , UpperCAmelCase=0 , UpperCAmelCase=-1 ) ->int: """simple docstring""" if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input A = tf.shape(UpperCAmelCase ) A = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) A = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(UpperCAmelCase , UpperCAmelCase ) def __a ( UpperCAmelCase ) ->tf.Tensor: """simple docstring""" if not isinstance(UpperCAmelCase , tf.Tensor ): A = tf.convert_to_tensor(UpperCAmelCase ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: A = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: A = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) A = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = "input_ids" ) ->None: """simple docstring""" tf.debugging.assert_less( UpperCAmelCase , tf.cast(UpperCAmelCase , dtype=tensor.dtype ) , message=( f"""The maximum value of {tensor_name} ({tf.math.reduce_max(UpperCAmelCase )}) must be smaller than the embedding """ f"""layer's input dimension ({embed_dim}). The likely cause is some problem at tokenization time.""" ) , ) def __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->Optional[Any]: """simple docstring""" A = 64512 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. A = [x for x in data if len(UpperCAmelCase ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( """The following attributes cannot be saved to HDF5 file because """ f"""they are larger than {HDF5_OBJECT_HEADER_LIMIT} """ f"""bytes: {bad_attributes}""" ) A = np.asarray(UpperCAmelCase ) A = 1 A = np.array_split(UpperCAmelCase , UpperCAmelCase ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 A = np.array_split(UpperCAmelCase , UpperCAmelCase ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(UpperCAmelCase ): A = chunk_data else: A = data def __a ( UpperCAmelCase , UpperCAmelCase ) ->int: """simple docstring""" if name in group.attrs: A = [n.decode("""utf8""" ) if hasattr(UpperCAmelCase , """decode""" ) else n for n in group.attrs[name]] else: A = [] A = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("""utf8""" ) if hasattr(UpperCAmelCase , """decode""" ) else n for n in group.attrs["""%s%d""" % (name, chunk_id)]] ) chunk_id += 1 return data def __a ( UpperCAmelCase ) ->Optional[Any]: """simple docstring""" def _expand_single_ad_tensor(UpperCAmelCase ): if isinstance(UpperCAmelCase , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(UpperCAmelCase , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , UpperCAmelCase )
337
0
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = """▁""" lowerCamelCase__ = { """vocab_file""": """vocab.json""", """spm_file""": """sentencepiece.bpe.model""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCamelCase__ = { """vocab_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json""", }, """spm_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model""", }, """tokenizer_config_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json""", }, } lowerCamelCase__ = { """facebook/m2m100_418M""": 1024, } # fmt: off lowerCamelCase__ = { """m2m100""": ["""af""", """am""", """ar""", """ast""", """az""", """ba""", """be""", """bg""", """bn""", """br""", """bs""", """ca""", """ceb""", """cs""", """cy""", """da""", """de""", """el""", """en""", """es""", """et""", """fa""", """ff""", """fi""", """fr""", """fy""", """ga""", """gd""", """gl""", """gu""", """ha""", """he""", """hi""", """hr""", """ht""", """hu""", """hy""", """id""", """ig""", """ilo""", """is""", """it""", """ja""", """jv""", """ka""", """kk""", """km""", """kn""", """ko""", """lb""", """lg""", """ln""", """lo""", """lt""", """lv""", """mg""", """mk""", """ml""", """mn""", """mr""", """ms""", """my""", """ne""", """nl""", """no""", """ns""", """oc""", """or""", """pa""", """pl""", """ps""", """pt""", """ro""", """ru""", """sd""", """si""", """sk""", """sl""", """so""", """sq""", """sr""", """ss""", """su""", """sv""", """sw""", """ta""", """th""", """tl""", """tn""", """tr""", """uk""", """ur""", """uz""", """vi""", """wo""", """xh""", """yi""", """yo""", """zh""", """zu"""], """wmt21""": ["""en""", """ha""", """is""", """ja""", """cs""", """ru""", """zh""", """de"""] } class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = ['input_ids', 'attention_mask'] lowercase = [] lowercase = [] def __init__( self : Optional[int] , a : Union[str, Any] , a : Dict , a : List[Any]=None , a : int=None , a : Optional[Any]="<s>" , a : Optional[Any]="</s>" , a : List[Any]="</s>" , a : Optional[Any]="<pad>" , a : Union[str, Any]="<unk>" , a : Tuple="m2m100" , a : Optional[Dict[str, Any]] = None , a : Tuple=8 , **a : Union[str, Any] , ): '''simple docstring''' lowerCAmelCase__ : Any = {} if sp_model_kwargs is None else sp_model_kwargs lowerCAmelCase__ : Optional[Any] = language_codes lowerCAmelCase__ : int = FAIRSEQ_LANGUAGE_CODES[language_codes] lowerCAmelCase__ : List[str] = {lang_code: f'''__{lang_code}__''' for lang_code in fairseq_language_code} lowerCAmelCase__ : Optional[Any] = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(a ) for lang_code in fairseq_language_code if self.get_lang_token(a ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=a , tgt_lang=a , bos_token=a , eos_token=a , sep_token=a , unk_token=a , pad_token=a , language_codes=a , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=a , **a , ) lowerCAmelCase__ : str = vocab_file lowerCAmelCase__ : List[Any] = load_json(a ) lowerCAmelCase__ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase__ : Optional[Any] = spm_file lowerCAmelCase__ : List[str] = load_spm(a , self.sp_model_kwargs ) lowerCAmelCase__ : List[str] = len(self.encoder ) lowerCAmelCase__ : int = { self.get_lang_token(a ): self.encoder_size + i for i, lang_code in enumerate(a ) } lowerCAmelCase__ : Tuple = {lang_code: self.encoder_size + i for i, lang_code in enumerate(a )} lowerCAmelCase__ : Optional[Any] = {v: k for k, v in self.lang_token_to_id.items()} lowerCAmelCase__ : str = src_lang if src_lang is not None else 'en' lowerCAmelCase__ : Any = tgt_lang lowerCAmelCase__ : int = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) lowerCAmelCase__ : Union[str, Any] = num_madeup_words @property def _lowerCamelCase ( self : str ): '''simple docstring''' return len(self.encoder ) + len(self.lang_token_to_id ) @property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return self._src_lang @src_lang.setter def _lowerCamelCase ( self : Any , a : str ): '''simple docstring''' lowerCAmelCase__ : List[Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowerCamelCase ( self : str , a : str ): '''simple docstring''' return self.sp_model.encode(a , out_type=a ) def _lowerCamelCase ( self : int , a : Dict ): '''simple docstring''' if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(a , self.encoder[self.unk_token] ) def _lowerCamelCase ( self : int , a : int ): '''simple docstring''' if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(a , self.unk_token ) def _lowerCamelCase ( self : List[Any] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : List[str] = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(a ) + token lowerCAmelCase__ : str = [] else: current_sub_tokens.append(a ) out_string += self.sp_model.decode(a ) return out_string.strip() def _lowerCamelCase ( self : List[str] , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) lowerCAmelCase__ : Tuple = [1] * len(self.prefix_tokens ) lowerCAmelCase__ : Dict = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(a )) + suffix_ones return prefix_ones + ([0] * len(a )) + ([0] * len(a )) + suffix_ones def _lowerCamelCase ( self : List[str] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.__dict__.copy() lowerCAmelCase__ : List[Any] = None return state def __setstate__( self : List[Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCAmelCase__ : Optional[Any] = {} lowerCAmelCase__ : List[Any] = load_spm(self.spm_file , self.sp_model_kwargs ) def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' lowerCAmelCase__ : Any = Path(a ) if not save_dir.is_dir(): raise OSError(f'''{save_directory} should be a directory''' ) lowerCAmelCase__ : Optional[int] = save_dir / ( (filename_prefix + '-' if filename_prefix else '') + self.vocab_files_names['vocab_file'] ) lowerCAmelCase__ : Optional[Any] = save_dir / ( (filename_prefix + '-' if filename_prefix else '') + self.vocab_files_names['spm_file'] ) save_json(self.encoder , a ) if os.path.abspath(self.spm_file ) != os.path.abspath(a ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , a ) elif not os.path.isfile(self.spm_file ): with open(a , 'wb' ) as fi: lowerCAmelCase__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(a ) return (str(a ), str(a )) def _lowerCamelCase ( self : Any , a : List[str] , a : str = "en" , a : Optional[List[str]] = None , a : str = "ro" , **a : Union[str, Any] , ): '''simple docstring''' lowerCAmelCase__ : Tuple = src_lang lowerCAmelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(a , a , **a ) def _lowerCamelCase ( self : Dict , a : List[str] , a : Optional[str] , a : Optional[str] , **a : str ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCAmelCase__ : Dict = src_lang lowerCAmelCase__ : Any = self(a , add_special_tokens=a , **a ) lowerCAmelCase__ : Union[str, Any] = self.get_lang_id(a ) lowerCAmelCase__ : List[Any] = tgt_lang_id return inputs def _lowerCamelCase ( self : str ): '''simple docstring''' self.set_src_lang_special_tokens(self.src_lang ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowerCamelCase ( self : Any , a : str ): '''simple docstring''' lowerCAmelCase__ : Dict = self.get_lang_token(a ) lowerCAmelCase__ : Tuple = self.lang_token_to_id[lang_token] lowerCAmelCase__ : Optional[int] = [self.cur_lang_id] lowerCAmelCase__ : Any = [self.eos_token_id] def _lowerCamelCase ( self : List[str] , a : str ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.get_lang_token(a ) lowerCAmelCase__ : Optional[Any] = self.lang_token_to_id[lang_token] lowerCAmelCase__ : List[str] = [self.cur_lang_id] lowerCAmelCase__ : str = [self.eos_token_id] def _lowerCamelCase ( self : List[Any] , a : str ): '''simple docstring''' return self.lang_code_to_token[lang] def _lowerCamelCase ( self : Union[str, Any] , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.get_lang_token(a ) return self.lang_token_to_id[lang_token] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> sentencepiece.SentencePieceProcessor: lowerCAmelCase__ : Dict = sentencepiece.SentencePieceProcessor(**SCREAMING_SNAKE_CASE_ ) spm.Load(str(SCREAMING_SNAKE_CASE_ ) ) return spm def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Union[Dict, List]: with open(SCREAMING_SNAKE_CASE_ , 'r' ) as f: return json.load(SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> None: with open(SCREAMING_SNAKE_CASE_ , 'w' ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , indent=2 )
212
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> np.array: lowerCAmelCase__ : Dict = F'''{sampling_rate}''' lowerCAmelCase__ : Any = '1' lowerCAmelCase__ : Optional[Any] = 'f32le' lowerCAmelCase__ : Any = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(SCREAMING_SNAKE_CASE_ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: lowerCAmelCase__ : List[Any] = ffmpeg_process.communicate(SCREAMING_SNAKE_CASE_ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error lowerCAmelCase__ : List[str] = output_stream[0] lowerCAmelCase__ : str = np.frombuffer(SCREAMING_SNAKE_CASE_ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = "f32le" , ) -> Dict: lowerCAmelCase__ : Optional[Any] = F'''{sampling_rate}''' lowerCAmelCase__ : Any = '1' if format_for_conversion == "s16le": lowerCAmelCase__ : Dict = 2 elif format_for_conversion == "f32le": lowerCAmelCase__ : List[str] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) lowerCAmelCase__ : Tuple = platform.system() if system == "Linux": lowerCAmelCase__ : str = 'alsa' lowerCAmelCase__ : str = 'default' elif system == "Darwin": lowerCAmelCase__ : Any = 'avfoundation' lowerCAmelCase__ : Tuple = ':0' elif system == "Windows": lowerCAmelCase__ : Any = 'dshow' lowerCAmelCase__ : int = 'default' lowerCAmelCase__ : Any = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] lowerCAmelCase__ : List[str] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample lowerCAmelCase__ : str = _ffmpeg_stream(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for item in iterator: yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = "f32le" , ) -> str: if stream_chunk_s is not None: lowerCAmelCase__ : Union[str, Any] = stream_chunk_s else: lowerCAmelCase__ : Tuple = chunk_length_s lowerCAmelCase__ : Any = ffmpeg_microphone(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , format_for_conversion=SCREAMING_SNAKE_CASE_ ) if format_for_conversion == "s16le": lowerCAmelCase__ : Optional[Any] = np.intaa lowerCAmelCase__ : Optional[Any] = 2 elif format_for_conversion == "f32le": lowerCAmelCase__ : Optional[Any] = np.floataa lowerCAmelCase__ : Optional[Any] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: lowerCAmelCase__ : Dict = chunk_length_s / 6 lowerCAmelCase__ : int = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(SCREAMING_SNAKE_CASE_ , (int, float) ): lowerCAmelCase__ : Dict = [stride_length_s, stride_length_s] lowerCAmelCase__ : Union[str, Any] = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample lowerCAmelCase__ : List[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample lowerCAmelCase__ : Any = datetime.datetime.now() lowerCAmelCase__ : Any = datetime.timedelta(seconds=SCREAMING_SNAKE_CASE_ ) for item in chunk_bytes_iter(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , stride=(stride_left, stride_right) , stream=SCREAMING_SNAKE_CASE_ ): # Put everything back in numpy scale lowerCAmelCase__ : Any = np.frombuffer(item['raw'] , dtype=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) lowerCAmelCase__ : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = False ) -> Optional[int]: lowerCAmelCase__ : Union[str, Any] = b'' lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = stride if stride_left + stride_right >= chunk_len: raise ValueError( F'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) lowerCAmelCase__ : List[str] = 0 for raw in iterator: acc += raw if stream and len(SCREAMING_SNAKE_CASE_ ) < chunk_len: lowerCAmelCase__ : Tuple = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(SCREAMING_SNAKE_CASE_ ) >= chunk_len: # We are flushing the accumulator lowerCAmelCase__ : Dict = (_stride_left, stride_right) lowerCAmelCase__ : Any = {'raw': acc[:chunk_len], 'stride': stride} if stream: lowerCAmelCase__ : Optional[int] = False yield item lowerCAmelCase__ : Optional[int] = stride_left lowerCAmelCase__ : Optional[int] = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(SCREAMING_SNAKE_CASE_ ) > stride_left: lowerCAmelCase__ : Tuple = {'raw': acc, 'stride': (_stride_left, 0)} if stream: lowerCAmelCase__ : Any = False yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: lowerCAmelCase__ : int = 2**24 # 16Mo try: with subprocess.Popen(SCREAMING_SNAKE_CASE_ , stdout=subprocess.PIPE , bufsize=SCREAMING_SNAKE_CASE_ ) as ffmpeg_process: while True: lowerCAmelCase__ : List[str] = ffmpeg_process.stdout.read(SCREAMING_SNAKE_CASE_ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
212
1
from __future__ import annotations import math import random from typing import Any class snake_case_: def __init__( self : Union[str, Any] ): lowerCAmelCase : list[Any] = [] lowerCAmelCase : int = 0 lowerCAmelCase : int = 0 def lowerCamelCase__ ( self : Dict ): return self.head == self.tail def lowerCamelCase__ ( self : Tuple , UpperCamelCase_ : Any ): self.data.append(UpperCamelCase_ ) lowerCAmelCase : Optional[int] = self.tail + 1 def lowerCamelCase__ ( self : Dict ): lowerCAmelCase : Tuple = self.data[self.head] lowerCAmelCase : Optional[int] = self.head + 1 return ret def lowerCamelCase__ ( self : Tuple ): return self.tail - self.head def lowerCamelCase__ ( self : Any ): print(self.data ) print('''**************''' ) print(self.data[self.head : self.tail] ) class snake_case_: def __init__( self : List[Any] , UpperCamelCase_ : Any ): lowerCAmelCase : List[str] = data lowerCAmelCase : MyNode | None = None lowerCAmelCase : MyNode | None = None lowerCAmelCase : int = 1 def lowerCamelCase__ ( self : int ): return self.data def lowerCamelCase__ ( self : Optional[Any] ): return self.left def lowerCamelCase__ ( self : List[str] ): return self.right def lowerCamelCase__ ( self : Optional[int] ): return self.height def lowerCamelCase__ ( self : Any , UpperCamelCase_ : Any ): lowerCAmelCase : Dict = data def lowerCamelCase__ ( self : Union[str, Any] , UpperCamelCase_ : MyNode | None ): lowerCAmelCase : Optional[Any] = node def lowerCamelCase__ ( self : Union[str, Any] , UpperCamelCase_ : MyNode | None ): lowerCAmelCase : str = node def lowerCamelCase__ ( self : Dict , UpperCamelCase_ : int ): lowerCAmelCase : Tuple = height def _snake_case ( _snake_case : MyNode | None ): if node is None: return 0 return node.get_height() def _snake_case ( _snake_case : int , _snake_case : int ): if a > b: return a return b def _snake_case ( _snake_case : MyNode ): print('''left rotation node:''' , node.get_data() ) lowerCAmelCase : Any = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(_snake_case ) lowerCAmelCase : Optional[int] = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) lowerCAmelCase : Optional[int] = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(_snake_case ) return ret def _snake_case ( _snake_case : MyNode ): print('''right rotation node:''' , node.get_data() ) lowerCAmelCase : List[Any] = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(_snake_case ) lowerCAmelCase : str = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) lowerCAmelCase : List[str] = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(_snake_case ) return ret def _snake_case ( _snake_case : MyNode ): lowerCAmelCase : List[str] = node.get_left() assert left_child is not None node.set_left(left_rotation(_snake_case ) ) return right_rotation(_snake_case ) def _snake_case ( _snake_case : MyNode ): lowerCAmelCase : List[str] = node.get_right() assert right_child is not None node.set_right(right_rotation(_snake_case ) ) return left_rotation(_snake_case ) def _snake_case ( _snake_case : MyNode | None , _snake_case : Any ): if node is None: return MyNode(_snake_case ) if data < node.get_data(): node.set_left(insert_node(node.get_left() , _snake_case ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected lowerCAmelCase : Dict = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child lowerCAmelCase : Optional[Any] = right_rotation(_snake_case ) else: lowerCAmelCase : Tuple = lr_rotation(_snake_case ) else: node.set_right(insert_node(node.get_right() , _snake_case ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: lowerCAmelCase : str = node.get_right() assert right_child is not None if data < right_child.get_data(): lowerCAmelCase : Any = rl_rotation(_snake_case ) else: lowerCAmelCase : Tuple = left_rotation(_snake_case ) lowerCAmelCase : Dict = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) return node def _snake_case ( _snake_case : MyNode ): while True: lowerCAmelCase : str = root.get_right() if right_child is None: break lowerCAmelCase : List[str] = right_child return root.get_data() def _snake_case ( _snake_case : MyNode ): while True: lowerCAmelCase : Dict = root.get_left() if left_child is None: break lowerCAmelCase : List[Any] = left_child return root.get_data() def _snake_case ( _snake_case : MyNode , _snake_case : Any ): lowerCAmelCase : Any = root.get_left() lowerCAmelCase : int = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: lowerCAmelCase : List[Any] = get_left_most(_snake_case ) root.set_data(_snake_case ) root.set_right(del_node(_snake_case , _snake_case ) ) elif left_child is not None: lowerCAmelCase : Dict = left_child elif right_child is not None: lowerCAmelCase : Tuple = right_child else: return None elif root.get_data() > data: if left_child is None: print('''No such data''' ) return root else: root.set_left(del_node(_snake_case , _snake_case ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(_snake_case , _snake_case ) ) if get_height(_snake_case ) - get_height(_snake_case ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): lowerCAmelCase : Any = left_rotation(_snake_case ) else: lowerCAmelCase : List[str] = rl_rotation(_snake_case ) elif get_height(_snake_case ) - get_height(_snake_case ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): lowerCAmelCase : List[str] = right_rotation(_snake_case ) else: lowerCAmelCase : Optional[int] = lr_rotation(_snake_case ) lowerCAmelCase : Union[str, Any] = my_max(get_height(root.get_right() ) , get_height(root.get_left() ) ) + 1 root.set_height(_snake_case ) return root class snake_case_: def __init__( self : Union[str, Any] ): lowerCAmelCase : MyNode | None = None def lowerCamelCase__ ( self : int ): return get_height(self.root ) def lowerCamelCase__ ( self : List[str] , UpperCamelCase_ : Any ): print('''insert:''' + str(UpperCamelCase_ ) ) lowerCAmelCase : str = insert_node(self.root , UpperCamelCase_ ) def lowerCamelCase__ ( self : List[Any] , UpperCamelCase_ : Any ): print('''delete:''' + str(UpperCamelCase_ ) ) if self.root is None: print('''Tree is empty!''' ) return lowerCAmelCase : int = del_node(self.root , UpperCamelCase_ ) def __str__( self : List[str] , ): # a level traversale, gives a more intuitive look on the tree lowerCAmelCase : List[str] = '''''' lowerCAmelCase : int = MyQueue() q.push(self.root ) lowerCAmelCase : Any = self.get_height() if layer == 0: return output lowerCAmelCase : Union[str, Any] = 0 while not q.is_empty(): lowerCAmelCase : int = q.pop() lowerCAmelCase : int = ''' ''' * int(math.pow(2 , layer - 1 ) ) output += space if node is None: output += "*" q.push(UpperCamelCase_ ) q.push(UpperCamelCase_ ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space lowerCAmelCase : Dict = cnt + 1 for i in range(1_0_0 ): if cnt == math.pow(2 , UpperCamelCase_ ) - 1: lowerCAmelCase : Union[str, Any] = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def _snake_case ( ): import doctest doctest.testmod() if __name__ == "__main__": _test() snake_case__ : int = AVLtree() snake_case__ : Optional[int] = list(range(10)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
362
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class snake_case_: def __init__( self : Dict , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : int=sys.maxsize ): lowerCAmelCase : Tuple = '''bilinear''' lowerCAmelCase : List[Any] = max_size lowerCAmelCase : Optional[int] = short_edge_length def __call__( self : Optional[int] , UpperCamelCase_ : Optional[int] ): lowerCAmelCase : Tuple = [] for img in imgs: lowerCAmelCase, lowerCAmelCase : List[str] = img.shape[:2] # later: provide list and randomly choose index for resize lowerCAmelCase : int = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img lowerCAmelCase : Optional[Any] = size * 1.0 / min(UpperCamelCase_ , UpperCamelCase_ ) if h < w: lowerCAmelCase, lowerCAmelCase : List[str] = size, scale * w else: lowerCAmelCase, lowerCAmelCase : int = scale * h, size if max(UpperCamelCase_ , UpperCamelCase_ ) > self.max_size: lowerCAmelCase : Union[str, Any] = self.max_size * 1.0 / max(UpperCamelCase_ , UpperCamelCase_ ) lowerCAmelCase : Tuple = newh * scale lowerCAmelCase : str = neww * scale lowerCAmelCase : Union[str, Any] = int(neww + 0.5 ) lowerCAmelCase : str = int(newh + 0.5 ) if img.dtype == np.uinta: lowerCAmelCase : Tuple = Image.fromarray(UpperCamelCase_ ) lowerCAmelCase : Optional[int] = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) lowerCAmelCase : Union[str, Any] = np.asarray(UpperCamelCase_ ) else: lowerCAmelCase : List[str] = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw lowerCAmelCase : Optional[int] = nn.functional.interpolate( UpperCamelCase_ , (newh, neww) , mode=self.interp_method , align_corners=UpperCamelCase_ ).squeeze(0 ) img_augs.append(UpperCamelCase_ ) return img_augs class snake_case_: def __init__( self : Tuple , UpperCamelCase_ : Any ): lowerCAmelCase : Any = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) lowerCAmelCase : List[Any] = cfg.INPUT.FORMAT lowerCAmelCase : Tuple = cfg.SIZE_DIVISIBILITY lowerCAmelCase : int = cfg.PAD_VALUE lowerCAmelCase : Union[str, Any] = cfg.INPUT.MAX_SIZE_TEST lowerCAmelCase : Union[str, Any] = cfg.MODEL.DEVICE lowerCAmelCase : Union[str, Any] = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) lowerCAmelCase : List[Any] = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) lowerCAmelCase : Optional[int] = lambda UpperCamelCase_ : (x - self.pixel_mean) / self.pixel_std def lowerCamelCase__ ( self : Optional[int] , UpperCamelCase_ : List[Any] ): lowerCAmelCase : Dict = tuple(max(UpperCamelCase_ ) for s in zip(*[img.shape for img in images] ) ) lowerCAmelCase : Dict = [im.shape[-2:] for im in images] lowerCAmelCase : Dict = [ nn.functional.pad( UpperCamelCase_ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(UpperCamelCase_ , UpperCamelCase_ ) ] return torch.stack(UpperCamelCase_ ), torch.tensor(UpperCamelCase_ ) def __call__( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[int]=False ): with torch.no_grad(): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): lowerCAmelCase : List[Any] = [images] if single_image: assert len(UpperCamelCase_ ) == 1 for i in range(len(UpperCamelCase_ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(UpperCamelCase_ , images.pop(UpperCamelCase_ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( UpperCamelCase_ , torch.as_tensor(img_tensorize(images.pop(UpperCamelCase_ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge lowerCAmelCase : Dict = torch.tensor([im.shape[:2] for im in images] ) lowerCAmelCase : str = self.aug(UpperCamelCase_ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic lowerCAmelCase : int = [self.normalizer(UpperCamelCase_ ) for x in images] # now pad them to do the following operations lowerCAmelCase, lowerCAmelCase : Optional[Any] = self.pad(UpperCamelCase_ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad lowerCAmelCase : Union[str, Any] = torch.true_divide(UpperCamelCase_ , UpperCamelCase_ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _snake_case ( _snake_case : str , _snake_case : List[Any] ): boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _snake_case ( _snake_case : Any , _snake_case : Tuple[int, int] ): assert torch.isfinite(_snake_case ).all(), "Box tensor contains infinite or NaN!" lowerCAmelCase, lowerCAmelCase : Optional[int] = box_size tensor[:, 0].clamp_(min=0 , max=_snake_case ) tensor[:, 1].clamp_(min=0 , max=_snake_case ) tensor[:, 2].clamp_(min=0 , max=_snake_case ) tensor[:, 3].clamp_(min=0 , max=_snake_case )
314
0
import torch def __SCREAMING_SNAKE_CASE ( ): '''simple docstring''' if torch.cuda.is_available(): _UpperCAmelCase = torch.cuda.device_count() else: _UpperCAmelCase = 0 print(f"""Successfully ran on {num_gpus} GPUs""" ) if __name__ == "__main__": main()
133
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = { "allenai/longformer-base-4096": "https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json", "allenai/longformer-large-4096": "https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json", "allenai/longformer-large-4096-finetuned-triviaqa": ( "https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json" ), "allenai/longformer-base-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json" ), "allenai/longformer-large-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json" ), } class lowercase ( _UpperCAmelCase ): _SCREAMING_SNAKE_CASE = 'longformer' def __init__( self , lowercase = 512 , lowercase = 2 , lowercase = 1 , lowercase = 0 , lowercase = 2 , lowercase = 30_522 , lowercase = 768 , lowercase = 12 , lowercase = 12 , lowercase = 3_072 , lowercase = "gelu" , lowercase = 0.1 , lowercase = 0.1 , lowercase = 512 , lowercase = 2 , lowercase = 0.02 , lowercase = 1e-12 , lowercase = False , **lowercase , ) -> Optional[int]: super().__init__(pad_token_id=lowercase , **lowercase ) lowerCAmelCase = attention_window lowerCAmelCase = sep_token_id lowerCAmelCase = bos_token_id lowerCAmelCase = eos_token_id lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = hidden_act lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = initializer_range lowerCAmelCase = layer_norm_eps lowerCAmelCase = onnx_export class lowercase ( _UpperCAmelCase ): def __init__( self , lowercase , lowercase = "default" , lowercase = None ) -> Tuple: super().__init__(lowercase , lowercase , lowercase ) lowerCAmelCase = True @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""global_attention_mask""", dynamic_axis), ] ) @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: lowerCAmelCase = super().outputs if self.task == "default": lowerCAmelCase = {0: """batch"""} return outputs @property def _snake_case ( self ) -> float: return 1e-4 @property def _snake_case ( self ) -> int: # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 14 ) def _snake_case ( self , lowercase , lowercase = -1 , lowercase = -1 , lowercase = False , lowercase = None , ) -> Mapping[str, Any]: lowerCAmelCase = super().generate_dummy_inputs( preprocessor=lowercase , batch_size=lowercase , seq_length=lowercase , is_pair=lowercase , framework=lowercase ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly lowerCAmelCase = torch.zeros_like(inputs["""input_ids"""] ) # make every second token global lowerCAmelCase = 1 return inputs
46
0
"""simple docstring""" def _a ( _SCREAMING_SNAKE_CASE ) -> bool: snake_case_ = set() # To detect a back edge, keep track of vertices currently in the recursion stack snake_case_ = set() return any( node not in visited and depth_first_search(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for node in graph ) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> bool: visited.add(_SCREAMING_SNAKE_CASE ) rec_stk.add(_SCREAMING_SNAKE_CASE ) for node in graph[vertex]: if node not in visited: if depth_first_search(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(_SCREAMING_SNAKE_CASE ) return False if __name__ == "__main__": from doctest import testmod testmod()
358
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> int: snake_case_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """vit.embeddings.cls_token"""), ("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" snake_case_ = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> List[Any]: for i in range(config.num_hidden_layers ): if base_model: snake_case_ = """""" else: snake_case_ = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case_ = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) snake_case_ = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case_ = in_proj_weight[ : config.hidden_size, : ] snake_case_ = in_proj_bias[: config.hidden_size] snake_case_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case_ = in_proj_weight[ -config.hidden_size :, : ] snake_case_ = in_proj_bias[-config.hidden_size :] def _a ( _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: snake_case_ = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: snake_case_ = dct.pop(_SCREAMING_SNAKE_CASE ) snake_case_ = val def _a ( ) -> Dict: snake_case_ = """http://images.cocodataset.org/val2017/000000039769.jpg""" snake_case_ = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ) return im @torch.no_grad() def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True ) -> Any: snake_case_ = ViTConfig() # patch_size if model_name[-1] == "8": snake_case_ = 8 # set labels if required if not base_model: snake_case_ = 1_000 snake_case_ = """huggingface/label-files""" snake_case_ = """imagenet-1k-id2label.json""" snake_case_ = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) , """r""" ) ) snake_case_ = {int(_SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: snake_case_ = 384 snake_case_ = 1_536 snake_case_ = 12 snake_case_ = 6 # load original model from torch hub snake_case_ = torch.hub.load("""facebookresearch/dino:main""" , _SCREAMING_SNAKE_CASE ) original_model.eval() # load state_dict of original model, remove and rename some keys snake_case_ = original_model.state_dict() if base_model: remove_classification_head_(_SCREAMING_SNAKE_CASE ) snake_case_ = create_rename_keys(_SCREAMING_SNAKE_CASE , base_model=_SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) read_in_q_k_v(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # load HuggingFace model if base_model: snake_case_ = ViTModel(_SCREAMING_SNAKE_CASE , add_pooling_layer=_SCREAMING_SNAKE_CASE ).eval() else: snake_case_ = ViTForImageClassification(_SCREAMING_SNAKE_CASE ).eval() model.load_state_dict(_SCREAMING_SNAKE_CASE ) # Check outputs on an image, prepared by ViTImageProcessor snake_case_ = ViTImageProcessor() snake_case_ = image_processor(images=prepare_img() , return_tensors="""pt""" ) snake_case_ = encoding["""pixel_values"""] snake_case_ = model(_SCREAMING_SNAKE_CASE ) if base_model: snake_case_ = original_model(_SCREAMING_SNAKE_CASE ) assert torch.allclose(_SCREAMING_SNAKE_CASE , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: snake_case_ = original_model(_SCREAMING_SNAKE_CASE ) assert logits.shape == outputs.logits.shape assert torch.allclose(_SCREAMING_SNAKE_CASE , outputs.logits , atol=1E-3 ) Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
233
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class A ( unittest.TestCase ): def lowercase_ (self : List[Any] ) -> Any: """simple docstring""" UpperCAmelCase__ = "ZinengTang/tvlt-base" UpperCAmelCase__ = tempfile.mkdtemp() def lowercase_ (self : Optional[Any] , **__UpperCAmelCase : Optional[Any] ) -> Tuple: """simple docstring""" return TvltImageProcessor.from_pretrained(self.checkpoint , **_UpperCamelCase ) def lowercase_ (self : Union[str, Any] , **__UpperCAmelCase : Optional[int] ) -> Any: """simple docstring""" return TvltFeatureExtractor.from_pretrained(self.checkpoint , **_UpperCamelCase ) def lowercase_ (self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowercase_ (self : List[str] ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = self.get_feature_extractor() UpperCAmelCase__ = TvltProcessor(image_processor=_UpperCamelCase , feature_extractor=_UpperCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , _UpperCamelCase ) self.assertIsInstance(processor.image_processor , _UpperCamelCase ) def lowercase_ (self : List[Any] ) -> Any: """simple docstring""" UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = self.get_feature_extractor() UpperCAmelCase__ = TvltProcessor(image_processor=_UpperCamelCase , feature_extractor=_UpperCamelCase ) UpperCAmelCase__ = np.ones([1_2_0_0_0] ) UpperCAmelCase__ = feature_extractor(_UpperCamelCase , return_tensors="np" ) UpperCAmelCase__ = processor(audio=_UpperCamelCase , return_tensors="np" ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowercase_ (self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = self.get_feature_extractor() UpperCAmelCase__ = TvltProcessor(image_processor=_UpperCamelCase , feature_extractor=_UpperCamelCase ) UpperCAmelCase__ = np.ones([3, 2_2_4, 2_2_4] ) UpperCAmelCase__ = image_processor(_UpperCamelCase , return_tensors="np" ) UpperCAmelCase__ = processor(images=_UpperCamelCase , return_tensors="np" ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowercase_ (self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = self.get_feature_extractor() UpperCAmelCase__ = TvltProcessor(image_processor=_UpperCamelCase , feature_extractor=_UpperCamelCase ) UpperCAmelCase__ = np.ones([1_2_0_0_0] ) UpperCAmelCase__ = np.ones([3, 2_2_4, 2_2_4] ) UpperCAmelCase__ = processor(audio=_UpperCamelCase , images=_UpperCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["audio_values", "audio_mask", "pixel_values", "pixel_mask"] ) # test if it raises when no input is passed with pytest.raises(_UpperCamelCase ): processor() def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.get_image_processor() UpperCAmelCase__ = self.get_feature_extractor() UpperCAmelCase__ = TvltProcessor(image_processor=_UpperCamelCase , feature_extractor=_UpperCamelCase ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg="`processor` and `image_processor`+`feature_extractor` model input names do not match" , )
65
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass snake_case__ : Union[str, Any] = (3, 9, -11, 0, 7, 5, 1, -1) snake_case__ : int = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class A_ : lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 class A_ : def __init__(self :Dict , _UpperCamelCase :Iterable[int] )-> None: __A = None for i in sorted(_UpperCamelCase , reverse=_UpperCamelCase ): __A = Node(_UpperCamelCase , self.head ) def __iter__(self :List[str] )-> Iterator[int]: __A = self.head while node: yield node.data __A = node.next_node def __len__(self :Union[str, Any] )-> int: return sum(1 for _ in self ) def __str__(self :List[Any] )-> str: return " -> ".join([str(_UpperCamelCase ) for node in self] ) def _a ( lowerCamelCase: SortedLinkedList , lowerCamelCase: SortedLinkedList ) -> SortedLinkedList: '''simple docstring''' return SortedLinkedList(list(lowerCamelCase ) + list(lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() snake_case__ : Any = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
117
0
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("""Googling.....""") __a: List[Any] = """https://www.google.com/search?q=""" + """ """.join(sys.argv[1:]) __a: int = requests.get(url, headers={"""UserAgent""": UserAgent().random}) # res.raise_for_status() with open("""project1a.html""", """wb""") as out_file: # only for knowing the class for data in res.iter_content(1_00_00): out_file.write(data) __a: Any = BeautifulSoup(res.text, """html.parser""") __a: Union[str, Any] = list(soup.select(""".eZt8xd"""))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get("""href""")) else: webbrowser.open(F'https://google.com{link.get("href")}')
356
'''simple docstring''' import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase ( a__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = OpenAIGPTTokenizer SCREAMING_SNAKE_CASE = OpenAIGPTTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _lowerCAmelCase( self ) -> Optional[int]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase__ : Optional[Any] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] lowercase__ : Optional[int] = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) lowercase__ : Union[str, Any] = ['''#version: 0.2''', '''l o''', '''lo w''', '''e r</w>''', ''''''] lowercase__ : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase__ : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' ) as fp: fp.write(json.dumps(__lowerCAmelCase ) ) with open(self.merges_file , '''w''' ) as fp: fp.write('''\n'''.join(__lowerCAmelCase ) ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> Dict: return "lower newer", "lower newer" def _lowerCAmelCase( self ) -> List[Any]: lowercase__ : int = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) lowercase__ : List[Any] = '''lower''' lowercase__ : Any = ['''low''', '''er</w>'''] lowercase__ : Union[str, Any] = tokenizer.tokenize(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) lowercase__ : Tuple = tokens + ['''<unk>'''] lowercase__ : Optional[Any] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) , __lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase=15 ) -> Optional[int]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowercase__ : List[str] = self.rust_tokenizer_class.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase ) # Simple input lowercase__ : List[str] = '''This is a simple input''' lowercase__ : Any = ['''This is a simple input 1''', '''This is a simple input 2'''] lowercase__ : str = ('''This is a simple input''', '''This is a pair''') lowercase__ : str = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(__lowerCAmelCase , tokenizer_r.encode , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Simple input self.assertRaises(__lowerCAmelCase , tokenizer_r.encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Simple input self.assertRaises( __lowerCAmelCase , tokenizer_r.batch_encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' , ) # Pair input self.assertRaises(__lowerCAmelCase , tokenizer_r.encode , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Pair input self.assertRaises(__lowerCAmelCase , tokenizer_r.encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Pair input self.assertRaises( __lowerCAmelCase , tokenizer_r.batch_encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' , ) def _lowerCAmelCase( self ) -> Tuple: pass @require_ftfy @require_spacy @require_tokenizers class UpperCAmelCase ( a__ ): '''simple docstring''' pass
214
0
'''simple docstring''' from __future__ import annotations import requests def lowercase_ ( _lowercase ) -> dict: '''simple docstring''' lowerCamelCase_ : Optional[int] = F"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(_lowercase ).json() def lowercase_ ( _lowercase = 10 ) -> list[dict]: '''simple docstring''' lowerCamelCase_ : Union[str, Any] = '''https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty''' lowerCamelCase_ : Tuple = requests.get(_lowercase ).json()[:max_stories] return [get_hackernews_story(_lowercase ) for story_id in story_ids] def lowercase_ ( _lowercase = 10 ) -> str: '''simple docstring''' lowerCamelCase_ : Optional[Any] = hackernews_top_stories(_lowercase ) return "\n".join('''* [{title}]({url})'''.format(**_lowercase ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
318
'''simple docstring''' import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor __lowercase : Dict = logging.get_logger(__name__) class __lowercase ( _lowercase ): def __init__(self , *A , **A ): warnings.warn( '''The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use OwlViTImageProcessor instead.''' , A , ) super().__init__(*A , **A )
318
1
__UpperCAmelCase = { '''a''': '''AAAAA''', '''b''': '''AAAAB''', '''c''': '''AAABA''', '''d''': '''AAABB''', '''e''': '''AABAA''', '''f''': '''AABAB''', '''g''': '''AABBA''', '''h''': '''AABBB''', '''i''': '''ABAAA''', '''j''': '''BBBAA''', '''k''': '''ABAAB''', '''l''': '''ABABA''', '''m''': '''ABABB''', '''n''': '''ABBAA''', '''o''': '''ABBAB''', '''p''': '''ABBBA''', '''q''': '''ABBBB''', '''r''': '''BAAAA''', '''s''': '''BAAAB''', '''t''': '''BAABA''', '''u''': '''BAABB''', '''v''': '''BBBAB''', '''w''': '''BABAA''', '''x''': '''BABAB''', '''y''': '''BABBA''', '''z''': '''BABBB''', ''' ''': ''' ''', } __UpperCAmelCase = {value: key for key, value in encode_dict.items()} def UpperCamelCase ( snake_case__ : str ) -> str: UpperCamelCase : List[Any] = '' for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception('encode() accepts only letters of the alphabet and spaces' ) return encoded def UpperCamelCase ( snake_case__ : str ) -> str: if set(lowerCAmelCase__ ) - {"A", "B", " "} != set(): raise Exception('decode() accepts only \'A\', \'B\' and spaces' ) UpperCamelCase : str = '' for word in coded.split(): while len(lowerCAmelCase__ ) != 0: decoded += decode_dict[word[:5]] UpperCamelCase : Optional[Any] = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
361
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase ( snake_case__ : List[str] , snake_case__ : List[str]=False ) -> Optional[Any]: try: UpperCamelCase : List[Any] = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCamelCase : List[Any] = default else: # KEY is set, convert it to True or False. try: UpperCamelCase : Optional[Any] = strtobool(snake_case__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F"""If set, {key} must be yes or no.""" ) return _value __UpperCAmelCase = parse_flag_from_env('''RUN_SLOW''', default=False) def UpperCamelCase ( snake_case__ : int ) -> str: return unittest.skip('Test was skipped' )(snake_case__ ) def UpperCamelCase ( snake_case__ : List[Any] ) -> Optional[Any]: return unittest.skipUnless(_run_slow_tests , 'test is slow' )(snake_case__ ) def UpperCamelCase ( snake_case__ : List[Any] ) -> Dict: return unittest.skipUnless(not torch.cuda.is_available() , 'test requires only a CPU' )(snake_case__ ) def UpperCamelCase ( snake_case__ : List[Any] ) -> Dict: return unittest.skipUnless(torch.cuda.is_available() , 'test requires a GPU' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Tuple ) -> List[Any]: return unittest.skipUnless(is_xpu_available() , 'test requires a XPU' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[Any] ) -> List[Any]: return unittest.skipUnless(is_mps_available() , 'test requires a `mps` backend support in `torch`' )(snake_case__ ) def UpperCamelCase ( snake_case__ : List[str] ) -> Tuple: return unittest.skipUnless( is_transformers_available() and is_datasets_available() , 'test requires the Hugging Face suite' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[Any] ) -> List[Any]: return unittest.skipUnless(is_bnb_available() , 'test requires the bitsandbytes library' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Dict ) -> List[str]: return unittest.skipUnless(is_tpu_available() , 'test requires TPU' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[int] ) -> Optional[int]: return unittest.skipUnless(torch.cuda.device_count() == 1 , 'test requires a GPU' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[int] ) -> Dict: return unittest.skipUnless(torch.xpu.device_count() == 1 , 'test requires a XPU' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Tuple ) -> Any: return unittest.skipUnless(torch.cuda.device_count() > 1 , 'test requires multiple GPUs' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[Any] ) -> int: return unittest.skipUnless(torch.xpu.device_count() > 1 , 'test requires multiple XPUs' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[int] ) -> Any: return unittest.skipUnless(is_safetensors_available() , 'test requires safetensors' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Optional[int] ) -> Dict: return unittest.skipUnless(is_deepspeed_available() , 'test requires DeepSpeed' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Dict ) -> Optional[Any]: return unittest.skipUnless(is_torch_version('>=' , '1.12.0' ) , 'test requires torch version >= 1.12.0' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Dict=None , snake_case__ : Union[str, Any]=None ) -> Optional[Any]: if test_case is None: return partial(snake_case__ , version=snake_case__ ) return unittest.skipUnless(is_torch_version('>=' , snake_case__ ) , F"""test requires torch version >= {version}""" )(snake_case__ ) def UpperCamelCase ( snake_case__ : Dict ) -> Optional[Any]: return unittest.skipUnless(is_tensorboard_available() , 'test requires Tensorboard' )(snake_case__ ) def UpperCamelCase ( snake_case__ : str ) -> Tuple: return unittest.skipUnless(is_wandb_available() , 'test requires wandb' )(snake_case__ ) def UpperCamelCase ( snake_case__ : Any ) -> List[Any]: return unittest.skipUnless(is_comet_ml_available() , 'test requires comet_ml' )(snake_case__ ) __UpperCAmelCase = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase ( snake_case__ : Optional[Any] ) -> Optional[Any]: return unittest.skipUnless( _atleast_one_tracker_available , 'test requires at least one tracker to be available and for `comet_ml` to not be installed' , )(snake_case__ ) class lowerCAmelCase_ ( unittest.TestCase ): UpperCAmelCase__ : Optional[Any] = True @classmethod def snake_case_ ( cls ) -> Any: UpperCamelCase : Optional[Any] = tempfile.mkdtemp() @classmethod def snake_case_ ( cls ) -> Tuple: if os.path.exists(cls.tmpdir ): shutil.rmtree(cls.tmpdir ) def snake_case_ ( self ) -> str: if self.clear_on_setup: for path in Path(self.tmpdir ).glob('**/*' ): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(SCREAMING_SNAKE_CASE_ ) class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self ) -> Optional[int]: super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> Tuple: UpperCamelCase : Optional[Any] = mocks if isinstance(SCREAMING_SNAKE_CASE_, (tuple, list) ) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop ) def UpperCamelCase ( snake_case__ : Tuple ) -> Optional[int]: UpperCamelCase : Tuple = AcceleratorState() UpperCamelCase : Tuple = tensor[None].clone().to(state.device ) UpperCamelCase : str = gather(snake_case__ ).cpu() UpperCamelCase : Union[str, Any] = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , snake_case__ ): return False return True class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> int: UpperCamelCase : List[Any] = returncode UpperCamelCase : Tuple = stdout UpperCamelCase : Any = stderr async def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Optional[int] ) -> Union[str, Any]: while True: UpperCamelCase : List[str] = await stream.readline() if line: callback(snake_case__ ) else: break async def UpperCamelCase ( snake_case__ : Dict , snake_case__ : int=None , snake_case__ : Dict=None , snake_case__ : Any=None , snake_case__ : Optional[Any]=False , snake_case__ : Union[str, Any]=False ) -> _RunOutput: if echo: print('\nRunning: ' , ' '.join(snake_case__ ) ) UpperCamelCase : Dict = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=snake_case__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=snake_case__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCamelCase : Tuple = [] UpperCamelCase : int = [] def tee(snake_case__ : List[Any] , snake_case__ : Optional[int] , snake_case__ : Optional[int] , snake_case__ : List[Any]="" ): UpperCamelCase : Union[str, Any] = line.decode('utf-8' ).rstrip() sink.append(snake_case__ ) if not quiet: print(snake_case__ , snake_case__ , file=snake_case__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda snake_case__ : tee(snake_case__ , snake_case__ , sys.stdout , label='stdout:' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda snake_case__ : tee(snake_case__ , snake_case__ , sys.stderr , label='stderr:' ) ) ), ] , timeout=snake_case__ , ) return _RunOutput(await p.wait() , snake_case__ , snake_case__ ) def UpperCamelCase ( snake_case__ : Dict , snake_case__ : Any=None , snake_case__ : Tuple=None , snake_case__ : Any=180 , snake_case__ : Any=False , snake_case__ : Optional[int]=True ) -> _RunOutput: UpperCamelCase : int = asyncio.get_event_loop() UpperCamelCase : Tuple = loop.run_until_complete( _stream_subprocess(snake_case__ , env=snake_case__ , stdin=snake_case__ , timeout=snake_case__ , quiet=snake_case__ , echo=snake_case__ ) ) UpperCamelCase : str = ' '.join(snake_case__ ) if result.returncode > 0: UpperCamelCase : Union[str, Any] = '\n'.join(result.stderr ) raise RuntimeError( F"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" F"""The combined stderr from workers follows:\n{stderr}""" ) return result class lowerCAmelCase_ ( a__ ): pass def UpperCamelCase ( snake_case__ : List[str] , snake_case__ : str=False ) -> int: try: UpperCamelCase : Union[str, Any] = subprocess.check_output(snake_case__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(snake_case__ , 'decode' ): UpperCamelCase : Optional[int] = output.decode('utf-8' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F"""Command `{" ".join(snake_case__ )}` failed with the following error:\n\n{e.output.decode()}""" ) from e
103
0
'''simple docstring''' from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
93
'''simple docstring''' import argparse import copy def snake_case_ ( __SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" lowercase_ : List[Any] = {} with open(__SCREAMING_SNAKE_CASE ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: lowercase_ : Union[str, Any] = [] _list.append([line.split()[1], line.split()[2]] ) lowercase_ : str = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: lowercase_ : Optional[int] = [] _list.append([line.split()[0], line.split()[2]] ) lowercase_ : Dict = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def snake_case_ ( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Tuple ): """simple docstring""" with open(__SCREAMING_SNAKE_CASE ) as f: lowercase_ : List[str] = f.read(1 ) lowercase_ : Optional[int] = start_node lowercase_ : Any = [] lowercase_ : List[str] = start_node lowercase_ : Optional[Any] = 0 while visiting not in first_solution: lowercase_ : Any = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(__SCREAMING_SNAKE_CASE ) and k[0] not in first_solution: lowercase_ : List[Any] = k[1] lowercase_ : List[Any] = k[0] first_solution.append(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = distance_of_first_solution + int(__SCREAMING_SNAKE_CASE ) lowercase_ : int = best_node first_solution.append(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 lowercase_ : Optional[Any] = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def snake_case_ ( __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" lowercase_ : Tuple = [] for n in solution[1:-1]: lowercase_ : List[str] = solution.index(__SCREAMING_SNAKE_CASE ) for kn in solution[1:-1]: lowercase_ : Any = solution.index(__SCREAMING_SNAKE_CASE ) if n == kn: continue lowercase_ : Dict = copy.deepcopy(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = kn lowercase_ : List[Any] = n lowercase_ : str = 0 for k in _tmp[:-1]: lowercase_ : Tuple = _tmp[_tmp.index(__SCREAMING_SNAKE_CASE ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: lowercase_ : Optional[Any] = distance + int(i[1] ) _tmp.append(__SCREAMING_SNAKE_CASE ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) lowercase_ : Union[str, Any] = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda __SCREAMING_SNAKE_CASE : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def snake_case_ ( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[Any] ): """simple docstring""" lowercase_ : Optional[int] = 1 lowercase_ : List[str] = first_solution lowercase_ : Dict = [] lowercase_ : List[str] = distance_of_first_solution lowercase_ : Optional[Any] = solution while count <= iters: lowercase_ : int = find_neighborhood(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase_ : Any = 0 lowercase_ : Dict = neighborhood[index_of_best_solution] lowercase_ : Optional[Any] = len(__SCREAMING_SNAKE_CASE ) - 1 lowercase_ : Tuple = False while not found: lowercase_ : Optional[int] = 0 while i < len(__SCREAMING_SNAKE_CASE ): if best_solution[i] != solution[i]: lowercase_ : Tuple = best_solution[i] lowercase_ : Optional[int] = solution[i] break lowercase_ : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) lowercase_ : Tuple = True lowercase_ : Optional[int] = best_solution[:-1] lowercase_ : Optional[Any] = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: lowercase_ : Optional[Any] = cost lowercase_ : int = solution else: lowercase_ : Any = index_of_best_solution + 1 lowercase_ : Any = neighborhood[index_of_best_solution] if len(__SCREAMING_SNAKE_CASE ) >= size: tabu_list.pop(0 ) lowercase_ : List[Any] = count + 1 return best_solution_ever, best_cost def snake_case_ ( __SCREAMING_SNAKE_CASE : List[str]=None ): """simple docstring""" lowercase_ : Any = generate_neighbours(args.File ) lowercase_ , lowercase_ : Union[str, Any] = generate_first_solution( args.File , __SCREAMING_SNAKE_CASE ) lowercase_ , lowercase_ : Optional[int] = tabu_search( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , args.Iterations , args.Size , ) print(F'''Best solution: {best_sol}, with total distance: {best_cost}.''' ) if __name__ == "__main__": _lowercase : Any = argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
93
1
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class __SCREAMING_SNAKE_CASE : def __init__( self , _UpperCamelCase , _UpperCamelCase=13 , _UpperCamelCase=7 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=99 , _UpperCamelCase=32 , _UpperCamelCase=5 , _UpperCamelCase=4 , _UpperCamelCase=4 , _UpperCamelCase="gelu" , _UpperCamelCase=0.0 , _UpperCamelCase=0.1 , _UpperCamelCase=True , _UpperCamelCase=5_12 , _UpperCamelCase=16 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=3 , _UpperCamelCase=4 , _UpperCamelCase=None , ): """simple docstring""" lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = seq_length lowerCAmelCase__ = is_training lowerCAmelCase__ = use_input_mask lowerCAmelCase__ = use_token_type_ids lowerCAmelCase__ = use_labels lowerCAmelCase__ = vocab_size lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = intermediate_multiple_size lowerCAmelCase__ = hidden_act lowerCAmelCase__ = hidden_dropout lowerCAmelCase__ = attention_dropout lowerCAmelCase__ = weight_tying lowerCAmelCase__ = max_position_embeddings lowerCAmelCase__ = type_vocab_size lowerCAmelCase__ = type_sequence_label_size lowerCAmelCase__ = initializer_range lowerCAmelCase__ = num_labels lowerCAmelCase__ = num_choices lowerCAmelCase__ = scope def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ = None if self.use_input_mask: lowerCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase__ ( self ): """simple docstring""" return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_UpperCamelCase , initializer_range=self.initializer_range , ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ = True return config, input_ids, input_mask, token_labels def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = GPTNeoXJapaneseModel(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase__ = model(_UpperCamelCase , attention_mask=_UpperCamelCase ) lowerCAmelCase__ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = True lowerCAmelCase__ = GPTNeoXJapaneseModel(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase__ = model(_UpperCamelCase , attention_mask=_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = GPTNeoXJapaneseForCausalLM(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase__ = model(_UpperCamelCase , attention_mask=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = True lowerCAmelCase__ = GPTNeoXJapaneseForCausalLM(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() # first forward pass lowerCAmelCase__ = model(_UpperCamelCase , attention_mask=_UpperCamelCase , use_cache=_UpperCamelCase ) lowerCAmelCase__ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowerCAmelCase__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase__ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowerCAmelCase__ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCAmelCase__ = torch.cat([input_mask, next_mask] , dim=-1 ) lowerCAmelCase__ = model(_UpperCamelCase , attention_mask=_UpperCamelCase , output_hidden_states=_UpperCamelCase ) lowerCAmelCase__ = output_from_no_past['hidden_states'][0] lowerCAmelCase__ = model( _UpperCamelCase , attention_mask=_UpperCamelCase , past_key_values=_UpperCamelCase , output_hidden_states=_UpperCamelCase , )['hidden_states'][0] # select random slice lowerCAmelCase__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCAmelCase__ = output_from_no_past[:, -3:, random_slice_idx].detach() lowerCAmelCase__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_UpperCamelCase , _UpperCamelCase , atol=1E-3 ) ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = config_and_inputs lowerCAmelCase__ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( __lowercase , __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : Dict = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () _SCREAMING_SNAKE_CASE : Optional[int] = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () _SCREAMING_SNAKE_CASE : List[str] = ( {'''feature-extraction''': GPTNeoXJapaneseModel, '''text-generation''': GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE : Optional[int] = False _SCREAMING_SNAKE_CASE : Dict = False _SCREAMING_SNAKE_CASE : Dict = False _SCREAMING_SNAKE_CASE : Optional[Any] = False def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = GPTNeoXJapaneseModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=_UpperCamelCase , hidden_size=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" # This regression test was failing with PyTorch < 1.3 lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_decoder() lowerCAmelCase__ = None self.model_tester.create_and_check_model_as_decoder(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*_UpperCamelCase ) @slow def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = 'abeja/gpt-neox-japanese-2.7b' lowerCAmelCase__ = ['データサイエンティストとは、', '100年後に必要とされる会社は、', 'フルリモートの環境で働くために必要なことは、', '国境の長いトンネルを抜けると', '美味しい日本食といえば、'] lowerCAmelCase__ = [ 'データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。', '100年後に必要とされる会社は、「人」が中心の会社です。', 'フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。', '国境の長いトンネルを抜けると、そこは雪国だった。', '美味しい日本食といえば、やっぱりお寿司ですよね。', ] lowerCAmelCase__ = GPTNeoXJapaneseTokenizer.from_pretrained(_UpperCamelCase ) lowerCAmelCase__ = GPTNeoXJapaneseForCausalLM.from_pretrained(_UpperCamelCase ) lowerCAmelCase__ = [] for prompt in prompts: lowerCAmelCase__ = tokenizer(_UpperCamelCase , return_tensors='pt' ).input_ids lowerCAmelCase__ = model.generate(_UpperCamelCase , max_length=50 ) lowerCAmelCase__ = tokenizer.batch_decode(_UpperCamelCase , skip_special_tokens=_UpperCamelCase ) predicted_outputs += generated_string self.assertListEqual(_UpperCamelCase , _UpperCamelCase )
122
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __SCREAMING_SNAKE_CASE ( __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : Tuple = BarthezTokenizer _SCREAMING_SNAKE_CASE : int = BarthezTokenizerFast _SCREAMING_SNAKE_CASE : Dict = True _SCREAMING_SNAKE_CASE : Tuple = True def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase__ = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_UpperCamelCase ) lowerCAmelCase__ = tokenizer def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = '<pad>' lowerCAmelCase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCamelCase ) , _UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCamelCase ) , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(_UpperCamelCase ) , 10_11_22 ) def UpperCamelCase__ ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] lowerCAmelCase__ = [0, 57, 30_18, 7_03_07, 91, 2] lowerCAmelCase__ = self.tokenizer( _UpperCamelCase , max_length=len(_UpperCamelCase ) , padding=_UpperCamelCase , truncation=_UpperCamelCase , return_tensors='pt' ) self.assertIsInstance(_UpperCamelCase , _UpperCamelCase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowerCAmelCase__ = batch.input_ids.tolist()[0] self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" if not self.test_rust_tokenizer: return lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = self.get_rust_tokenizer() lowerCAmelCase__ = 'I was born in 92000, and this is falsé.' lowerCAmelCase__ = tokenizer.tokenize(_UpperCamelCase ) lowerCAmelCase__ = rust_tokenizer.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) lowerCAmelCase__ = tokenizer.encode(_UpperCamelCase , add_special_tokens=_UpperCamelCase ) lowerCAmelCase__ = rust_tokenizer.encode(_UpperCamelCase , add_special_tokens=_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) lowerCAmelCase__ = self.get_rust_tokenizer() lowerCAmelCase__ = tokenizer.encode(_UpperCamelCase ) lowerCAmelCase__ = rust_tokenizer.encode(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) @slow def UpperCamelCase__ ( self ): """simple docstring""" # fmt: off lowerCAmelCase__ = {'input_ids': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowerCAmelCase__ = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_UpperCamelCase , model_name='moussaKam/mbarthez' , revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' , sequences=_UpperCamelCase , )
122
1
'''simple docstring''' import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase_ (_snake_case , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = CLIPTokenizer SCREAMING_SNAKE_CASE : Tuple = CLIPTokenizerFast SCREAMING_SNAKE_CASE : List[Any] = True SCREAMING_SNAKE_CASE : Dict = {} SCREAMING_SNAKE_CASE : Optional[int] = False def SCREAMING_SNAKE_CASE ( self : Dict ): super().setUp() # fmt: off __lowercase = ['l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on __lowercase = dict(zip(_UpperCamelCase ,range(len(_UpperCamelCase ) ) ) ) __lowercase = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>'] __lowercase = {'unk_token': '<unk>'} __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write(json.dumps(_UpperCamelCase ) + '''\n''' ) with open(self.merges_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(_UpperCamelCase ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,**lowercase__ : Union[str, Any] ): kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname ,**_UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,**lowercase__ : int ): kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname ,**_UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : List[str] ): __lowercase = 'lower newer' __lowercase = 'lower newer' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = CLIPTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) __lowercase = 'lower newer' __lowercase = ['lo', 'w', 'er</w>', 'n', 'e', 'w', 'er</w>'] __lowercase = tokenizer.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase ,_UpperCamelCase ) __lowercase = tokens + [tokenizer.unk_token] __lowercase = [1_0, 2, 1_6, 9, 3, 2, 1_6, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCamelCase ) ,_UpperCamelCase ) @require_ftfy def SCREAMING_SNAKE_CASE ( self : List[Any] ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.tokenizer_class.from_pretrained(_UpperCamelCase ,**_UpperCamelCase ) __lowercase = self.rust_tokenizer_class.from_pretrained(_UpperCamelCase ,**_UpperCamelCase ) __lowercase = 'A\n\'ll 11p223RF☆ho!!to?\'d\'d\'\'d of a cat to-$\'\'d.' __lowercase = tokenizer_s.tokenize(_UpperCamelCase ) __lowercase = tokenizer_r.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase ,_UpperCamelCase ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways __lowercase = 'xa\u0303y' + ' ' + 'x\xe3y' __lowercase = tokenizer_s.tokenize(_UpperCamelCase ) __lowercase = tokenizer_r.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase ,_UpperCamelCase ) # Test that the tokenization is identical on unicode of space type __lowercase = [ '\u0009', # (horizontal tab, '\t') '\u000B', # (vertical tab) '\u000C', # (form feed) '\u0020', # (space, ' ') '\u200E', # (left-to-right mark):w '\u200F', # (right-to-left mark) ] for unicode_seq in spaces_unicodes: __lowercase = tokenizer_s.tokenize(_UpperCamelCase ) __lowercase = tokenizer_r.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase ,_UpperCamelCase ) # Test that the tokenization is identical on unicode of line break type __lowercase = [ '\u000A', # (line feed, '\n') '\r\n', # (carriage return and line feed, '\r\n') '\u000D', # (carriage return, '\r') '\r', # (carriage return, '\r') '\u000D', # (carriage return, '\r') '\u2028', # (line separator) '\u2029', # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: __lowercase = tokenizer_s.tokenize(_UpperCamelCase ) __lowercase = tokenizer_r.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase ,_UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = 'hello' # `hello` is a token in the vocabulary of `pretrained_name` __lowercase = F"{text_of_1_token} {text_of_1_token}" __lowercase = self.rust_tokenizer_class.from_pretrained( _UpperCamelCase ,use_fast=_UpperCamelCase ,) __lowercase = tokenizer_r(_UpperCamelCase ,return_offsets_mapping=_UpperCamelCase ,add_special_tokens=_UpperCamelCase ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(_UpperCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(_UpperCamelCase ) + 1, len(_UpperCamelCase ) + 1 + len(_UpperCamelCase )) ,) __lowercase = F" {text}" __lowercase = self.rust_tokenizer_class.from_pretrained( _UpperCamelCase ,use_fast=_UpperCamelCase ,) __lowercase = tokenizer_r(_UpperCamelCase ,return_offsets_mapping=_UpperCamelCase ,add_special_tokens=_UpperCamelCase ) self.assertEqual(encoding.offset_mapping[0] ,(1, 1 + len(_UpperCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(_UpperCamelCase ) + 1, 1 + len(_UpperCamelCase ) + 1 + len(_UpperCamelCase )) ,) def SCREAMING_SNAKE_CASE ( self : List[Any] ): # Test related to the breaking change introduced in transformers v4.17.0 # We need to check that an error in raised when the user try to load a previous version of the tokenizer. with self.assertRaises(_UpperCamelCase ) as context: self.rust_tokenizer_class.from_pretrained('''robot-test/old-clip-tokenizer''' ) self.assertTrue( context.exception.args[0].startswith( '''The `backend_tokenizer` provided does not match the expected format.''' ) ) @require_ftfy def SCREAMING_SNAKE_CASE ( self : Tuple ): super().test_tokenization_python_rust_equals() def SCREAMING_SNAKE_CASE ( self : str ): # CLIP always lower cases letters pass
104
def lowercase__ ( __snake_case : list ): '''simple docstring''' for i in range(len(__snake_case ) - 1 , 0 , -1 ): UpperCAmelCase_ : Dict = False for j in range(__snake_case , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: UpperCAmelCase_ , UpperCAmelCase_ : Any = unsorted[j - 1], unsorted[j] UpperCAmelCase_ : int = True for j in range(__snake_case ): if unsorted[j] > unsorted[j + 1]: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = unsorted[j + 1], unsorted[j] UpperCAmelCase_ : Any = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = input('Enter numbers separated by a comma:\n').strip() __UpperCAmelCase = [int(item) for item in user_input.split(',')] print(F'{cocktail_shaker_sort(unsorted) = }')
29
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase : str = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = XGLMTokenizer __magic_name__ = XGLMTokenizerFast __magic_name__ = True __magic_name__ = True def a ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _lowerCAmelCase : List[str] = XGLMTokenizer(snake_case__ , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def a ( self ): '''simple docstring''' _lowerCAmelCase : str = '<pad>' _lowerCAmelCase : List[str] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(len(snake_case__ ) , 1008 ) def a ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1008 ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = XGLMTokenizer(snake_case__ , keep_accents=snake_case__ ) _lowerCAmelCase : Union[str, Any] = tokenizer.tokenize('This is a test' ) self.assertListEqual(snake_case__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _lowerCAmelCase : int = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) _lowerCAmelCase : List[str] = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _lowerCAmelCase : Tuple = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def a ( self ): '''simple docstring''' return XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) def a ( self ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(snake_case__ , f.name ) _lowerCAmelCase : int = XGLMTokenizer(f.name , keep_accents=snake_case__ ) _lowerCAmelCase : Any = pickle.dumps(snake_case__ ) pickle.loads(snake_case__ ) def a ( self ): '''simple docstring''' if not self.test_rust_tokenizer: return _lowerCAmelCase : List[str] = self.get_tokenizer() _lowerCAmelCase : Any = self.get_rust_tokenizer() _lowerCAmelCase : Tuple = 'I was born in 92000, and this is falsé.' _lowerCAmelCase : List[Any] = tokenizer.tokenize(snake_case__ ) _lowerCAmelCase : Dict = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : Union[str, Any] = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : Union[str, Any] = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : List[str] = self.get_rust_tokenizer() _lowerCAmelCase : int = tokenizer.encode(snake_case__ ) _lowerCAmelCase : List[Any] = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = 'Hello World!' _lowerCAmelCase : Optional[Any] = [2, 3_1227, 4447, 35] self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth' ) # fmt: off _lowerCAmelCase : int = [2, 1018, 67, 11, 1988, 2617, 5631, 278, 11, 3407, 48, 7_1630, 2_8085, 4, 3234, 157, 13, 6, 5, 6, 4, 3526, 768, 15, 659, 57, 298, 3983, 864, 129, 21, 6, 5, 1_3675, 377, 652, 7580, 1_0341, 155, 2817, 422, 1666, 7, 1674, 53, 113, 20_2277, 1_7892, 33, 60, 87, 4, 3234, 157, 61, 2667, 5_2376, 19, 88, 23, 735] # fmt: on self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = { 'input_ids': [[2, 10_8825, 1163, 15, 8_8010, 473, 1_5898, 157, 1_3672, 1857, 312, 8, 23_8021, 1163, 53, 1_3672, 1857, 312, 8, 5_3283, 18_2396, 8, 1_8566, 16, 3_6733, 4101, 8, 230, 24_4017, 12_2553, 7, 15, 13_2597, 4, 293, 1_2511, 7610, 4, 3414, 13_2597, 9, 4, 3_2361, 362, 4, 734, 2_8512, 3_2569, 18, 4, 3_2361, 2_6096, 1_4982, 73, 1_8715, 2_1433, 23_5261, 15, 492, 1_2427, 16, 53, 1_8715, 2_1433, 6_5454, 15, 2_3659, 563, 16, 278, 597, 2843, 595, 7931, 18_2396, 6_4186, 22, 886, 595, 13_2981, 53, 2_5540, 3449, 4_3982, 3_9901, 5951, 878, 330, 4, 2_7694, 8_0269, 312, 53, 6517, 1_1780, 611, 2_0408, 5], [2, 6, 13_2597, 67, 4_2897, 33, 592, 8, 16_3729, 2_5540, 361, 13_6997, 10_9514, 17_3230, 7, 501, 60, 10_2913, 196, 5631, 235, 6_3243, 473, 6, 23_1757, 74, 5277, 7905, 53, 3095, 3_7317, 22, 454, 18_3874, 5], [2, 268, 3_1298, 4_6530, 6, 13_2935, 4_3831, 7, 597, 32, 24, 3688, 9865, 5]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name='facebook/xglm-564M' , padding=snake_case__ , )
25
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( """The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion""" ) lowerCAmelCase : str = None lowerCAmelCase : Optional[int] = { """7B""": 1_10_08, """13B""": 1_38_24, """30B""": 1_79_20, """65B""": 2_20_16, """70B""": 2_86_72, } lowerCAmelCase : Optional[int] = { """7B""": 1, """7Bf""": 1, """13B""": 2, """13Bf""": 2, """30B""": 4, """65B""": 8, """70B""": 8, """70Bf""": 8, } def lowercase (_A , _A=1 , _A=2_5_6 ): """simple docstring""" return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def lowercase (_A ): """simple docstring""" with open(_A , 'r' ) as f: return json.load(_A ) def lowercase (_A , _A ): """simple docstring""" with open(_A , 'w' ) as f: json.dump(_A , _A ) def lowercase (_A , _A , _A , _A=True ): """simple docstring""" os.makedirs(_A , exist_ok=_A ) _lowerCAmelCase : Optional[Any] = os.path.join(_A , 'tmp' ) os.makedirs(_A , exist_ok=_A ) _lowerCAmelCase : Any = read_json(os.path.join(_A , 'params.json' ) ) _lowerCAmelCase : List[str] = NUM_SHARDS[model_size] _lowerCAmelCase : str = params['n_layers'] _lowerCAmelCase : Optional[int] = params['n_heads'] _lowerCAmelCase : int = n_heads // num_shards _lowerCAmelCase : Optional[int] = params['dim'] _lowerCAmelCase : Union[str, Any] = dim // n_heads _lowerCAmelCase : Union[str, Any] = 10_000.0 _lowerCAmelCase : str = 1.0 / (base ** (torch.arange(0 , _A , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: _lowerCAmelCase : Optional[Any] = params['n_kv_heads'] # for GQA / MQA _lowerCAmelCase : str = n_heads_per_shard // num_key_value_heads _lowerCAmelCase : Optional[int] = dim // num_key_value_heads else: # compatibility with other checkpoints _lowerCAmelCase : Union[str, Any] = n_heads _lowerCAmelCase : Any = n_heads_per_shard _lowerCAmelCase : Optional[Any] = dim # permute for sliced rotary def permute(_A , _A=n_heads , _A=dim , _A=dim ): return w.view(_A , dima // n_heads // 2 , 2 , _A ).transpose(1 , 2 ).reshape(_A , _A ) print(f'Fetching all parameters from the checkpoint at {input_base_path}.' ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) _lowerCAmelCase : List[Any] = torch.load(os.path.join(_A , 'consolidated.00.pth' ) , map_location='cpu' ) else: # Sharded _lowerCAmelCase : List[Any] = [ torch.load(os.path.join(_A , f'consolidated.{i:02d}.pth' ) , map_location='cpu' ) for i in range(_A ) ] _lowerCAmelCase : Tuple = 0 _lowerCAmelCase : Union[str, Any] = {'weight_map': {}} for layer_i in range(_A ): _lowerCAmelCase : List[str] = f'pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin' if model_size == "7B": # Unsharded _lowerCAmelCase : str = { f'model.layers.{layer_i}.self_attn.q_proj.weight': permute( loaded[f'layers.{layer_i}.attention.wq.weight'] ), f'model.layers.{layer_i}.self_attn.k_proj.weight': permute( loaded[f'layers.{layer_i}.attention.wk.weight'] ), f'model.layers.{layer_i}.self_attn.v_proj.weight': loaded[f'layers.{layer_i}.attention.wv.weight'], f'model.layers.{layer_i}.self_attn.o_proj.weight': loaded[f'layers.{layer_i}.attention.wo.weight'], f'model.layers.{layer_i}.mlp.gate_proj.weight': loaded[f'layers.{layer_i}.feed_forward.w1.weight'], f'model.layers.{layer_i}.mlp.down_proj.weight': loaded[f'layers.{layer_i}.feed_forward.w2.weight'], f'model.layers.{layer_i}.mlp.up_proj.weight': loaded[f'layers.{layer_i}.feed_forward.w3.weight'], f'model.layers.{layer_i}.input_layernorm.weight': loaded[f'layers.{layer_i}.attention_norm.weight'], f'model.layers.{layer_i}.post_attention_layernorm.weight': loaded[f'layers.{layer_i}.ffn_norm.weight'], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. _lowerCAmelCase : str = { f'model.layers.{layer_i}.input_layernorm.weight': loaded[0][ f'layers.{layer_i}.attention_norm.weight' ].clone(), f'model.layers.{layer_i}.post_attention_layernorm.weight': loaded[0][ f'layers.{layer_i}.ffn_norm.weight' ].clone(), } _lowerCAmelCase : List[str] = permute( torch.cat( [ loaded[i][f'layers.{layer_i}.attention.wq.weight'].view(_A , _A , _A ) for i in range(_A ) ] , dim=0 , ).reshape(_A , _A ) ) _lowerCAmelCase : Optional[int] = permute( torch.cat( [ loaded[i][f'layers.{layer_i}.attention.wk.weight'].view( _A , _A , _A ) for i in range(_A ) ] , dim=0 , ).reshape(_A , _A ) , _A , _A , _A , ) _lowerCAmelCase : Dict = torch.cat( [ loaded[i][f'layers.{layer_i}.attention.wv.weight'].view( _A , _A , _A ) for i in range(_A ) ] , dim=0 , ).reshape(_A , _A ) _lowerCAmelCase : Dict = torch.cat( [loaded[i][f'layers.{layer_i}.attention.wo.weight'] for i in range(_A )] , dim=1 ) _lowerCAmelCase : List[Any] = torch.cat( [loaded[i][f'layers.{layer_i}.feed_forward.w1.weight'] for i in range(_A )] , dim=0 ) _lowerCAmelCase : Tuple = torch.cat( [loaded[i][f'layers.{layer_i}.feed_forward.w2.weight'] for i in range(_A )] , dim=1 ) _lowerCAmelCase : List[Any] = torch.cat( [loaded[i][f'layers.{layer_i}.feed_forward.w3.weight'] for i in range(_A )] , dim=0 ) _lowerCAmelCase : int = inv_freq for k, v in state_dict.items(): _lowerCAmelCase : Optional[Any] = filename param_count += v.numel() torch.save(_A , os.path.join(_A , _A ) ) _lowerCAmelCase : Dict = f'pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin' if model_size == "7B": # Unsharded _lowerCAmelCase : List[str] = { 'model.embed_tokens.weight': loaded['tok_embeddings.weight'], 'model.norm.weight': loaded['norm.weight'], 'lm_head.weight': loaded['output.weight'], } else: _lowerCAmelCase : List[str] = { 'model.norm.weight': loaded[0]['norm.weight'], 'model.embed_tokens.weight': torch.cat( [loaded[i]['tok_embeddings.weight'] for i in range(_A )] , dim=1 ), 'lm_head.weight': torch.cat([loaded[i]['output.weight'] for i in range(_A )] , dim=0 ), } for k, v in state_dict.items(): _lowerCAmelCase : int = filename param_count += v.numel() torch.save(_A , os.path.join(_A , _A ) ) # Write configs _lowerCAmelCase : Tuple = {'total_size': param_count * 2} write_json(_A , os.path.join(_A , 'pytorch_model.bin.index.json' ) ) _lowerCAmelCase : Optional[int] = params['ffn_dim_multiplier'] if 'ffn_dim_multiplier' in params else 1 _lowerCAmelCase : int = params['multiple_of'] if 'multiple_of' in params else 2_5_6 _lowerCAmelCase : List[Any] = LlamaConfig( hidden_size=_A , intermediate_size=compute_intermediate_size(_A , _A , _A ) , num_attention_heads=params['n_heads'] , num_hidden_layers=params['n_layers'] , rms_norm_eps=params['norm_eps'] , num_key_value_heads=_A , ) config.save_pretrained(_A ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('Loading the checkpoint in a Llama model.' ) _lowerCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_A , torch_dtype=torch.floataa , low_cpu_mem_usage=_A ) # Avoid saving this as part of the config. del model.config._name_or_path print('Saving in the Transformers format.' ) model.save_pretrained(_A , safe_serialization=_A ) shutil.rmtree(_A ) def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f'Saving a {tokenizer_class.__name__} to {tokenizer_path}.' ) _lowerCAmelCase : List[Any] = tokenizer_class(_A ) tokenizer.save_pretrained(_A ) def lowercase (): """simple docstring""" _lowerCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '--input_dir' , help='Location of LLaMA weights, which contains tokenizer.model and model folders' , ) parser.add_argument( '--model_size' , choices=['7B', '7Bf', '13B', '13Bf', '30B', '65B', '70B', '70Bf', 'tokenizer_only'] , ) parser.add_argument( '--output_dir' , help='Location to write HF model and tokenizer' , ) parser.add_argument('--safe_serialization' , type=_A , help='Whether or not to save using `safetensors`.' ) _lowerCAmelCase : Any = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) _lowerCAmelCase : Dict = os.path.join(args.input_dir , 'tokenizer.model' ) write_tokenizer(args.output_dir , _A ) if __name__ == "__main__": main()
25
1
def UpperCAmelCase ( a_ , a_ ) -> bool: """simple docstring""" __A = len(a_ ) __A = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __A = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __A = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __A = subset[i - 1][j] if arr[i - 1] <= j: __A = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
15
import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def UpperCAmelCase ( a_ ) -> List[str]: """simple docstring""" return sum(param.float().sum() if "encoder.embeddings" not in key else 0 for key, param in state_dict.items() ) def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" __A = {} for key, value in state_dict.items(): if "text_encoder.embeddings" in key or "image_encoder.embeddings" in key: continue __A = key.replace("heads.cmd.mim_head.cls.predictions" , "mmm_image_head" ) __A = key.replace("heads.cmd.mlm_head.cls.predictions" , "mmm_text_head" ) __A = key.replace("heads.cmd.itm_head.cls" , "itm_head" ) __A = key.replace("heads.cmd.itm_head.pooler" , "itm_head.pooler" ) __A = key.replace("heads.cmd.clip_head.logit_scale" , "flava.logit_scale" ) __A = key.replace("heads.fairseq_mlm.cls.predictions" , "mlm_head" ) __A = key.replace("heads.imagenet.mim_head.cls.predictions" , "mim_head" ) __A = key.replace("mm_text_projection" , "flava.text_to_mm_projection" ) __A = key.replace("mm_image_projection" , "flava.image_to_mm_projection" ) __A = key.replace("image_encoder.module" , "flava.image_model" ) __A = key.replace("text_encoder.module" , "flava.text_model" ) __A = key.replace("mm_encoder.module.encoder.cls_token" , "flava.multimodal_model.cls_token" ) __A = key.replace("mm_encoder.module" , "flava.multimodal_model" ) __A = key.replace("text_projection" , "flava.text_projection" ) __A = key.replace("image_projection" , "flava.image_projection" ) __A = value.float() for key, value in codebook_state_dict.items(): __A = value return upgrade @torch.no_grad() def UpperCAmelCase ( a_ , a_ , a_ , a_=None ) -> Tuple: """simple docstring""" if config_path is not None: __A = FlavaConfig.from_pretrained(a_ ) else: __A = FlavaConfig() __A = FlavaForPreTraining(a_ ).eval() __A = convert_dalle_checkpoint(a_ , a_ , save_checkpoint=a_ ) if os.path.exists(a_ ): __A = torch.load(a_ , map_location="cpu" ) else: __A = torch.hub.load_state_dict_from_url(a_ , map_location="cpu" ) __A = upgrade_state_dict(a_ , a_ ) hf_model.load_state_dict(a_ ) __A = hf_model.state_dict() __A = count_parameters(a_ ) __A = count_parameters(a_ ) + count_parameters(a_ ) assert torch.allclose(a_ , a_ , atol=1E-3 ) hf_model.save_pretrained(a_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE :Any = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to flava checkpoint') parser.add_argument('--codebook_path', default=None, type=str, help='Path to flava codebook checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') SCREAMING_SNAKE_CASE :Optional[int] = parser.parse_args() convert_flava_checkpoint(args.checkpoint_path, args.codebook_path, args.pytorch_dump_folder_path, args.config_path)
15
1
from pathlib import Path import fire def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ ) -> Tuple: """simple docstring""" A__ = Path(lowercase_ ) A__ = Path(lowercase_ ) dest_dir.mkdir(exist_ok=lowercase_ ) for path in src_dir.iterdir(): A__ = [x.rstrip() for x in list(path.open().readlines() )][:n] A__ = dest_dir.joinpath(path.name ) print(lowercase_ ) dest_path.open('''w''' ).write('''\n'''.join(lowercase_ ) ) if __name__ == "__main__": fire.Fire(minify)
355
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowerCamelCase : Any = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : str = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys _lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
231
0
import os import pytest from attr import dataclass _lowerCamelCase : List[Any] = 'us-east-1' # defaults region @dataclass class UpperCamelCase_ : '''simple docstring''' UpperCAmelCase__ = 42 UpperCAmelCase__ = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" UpperCAmelCase__ = { """task_name""": """mnli""", """per_device_train_batch_size""": 16, """per_device_eval_batch_size""": 16, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 500, """save_steps""": 5500, } UpperCAmelCase__ = {**hyperparameters, """max_steps""": 1000} @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Union[str, Any]: '''simple docstring''' if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def SCREAMING_SNAKE_CASE ( self : Tuple) ->str: '''simple docstring''' return f"""{self.framework}-transfromers-test""" @property def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Optional[Any]: '''simple docstring''' return f"""./tests/sagemaker/scripts/{self.framework}""" @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Union[str, Any]: '''simple docstring''' if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='''class''' ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Dict: """simple docstring""" A__ = SageMakerTestEnvironment(framework=request.cls.framework )
14
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = CustomTokenizer pass
286
0
import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE : def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=[30, 30] , __lowerCAmelCase=2 , __lowerCAmelCase=3 , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=32 , __lowerCAmelCase=5 , __lowerCAmelCase=4 , __lowerCAmelCase=37 , __lowerCAmelCase="gelu" , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=10 , __lowerCAmelCase=0.02 , __lowerCAmelCase=3 , __lowerCAmelCase=None , __lowerCAmelCase=8 , __lowerCAmelCase=10 , ): UpperCamelCase__ = parent UpperCamelCase__ = batch_size UpperCamelCase__ = image_size UpperCamelCase__ = patch_size UpperCamelCase__ = num_channels UpperCamelCase__ = is_training UpperCamelCase__ = use_labels UpperCamelCase__ = hidden_size UpperCamelCase__ = num_hidden_layers UpperCamelCase__ = num_attention_heads UpperCamelCase__ = intermediate_size UpperCamelCase__ = hidden_act UpperCamelCase__ = hidden_dropout_prob UpperCamelCase__ = attention_probs_dropout_prob UpperCamelCase__ = type_sequence_label_size UpperCamelCase__ = initializer_range UpperCamelCase__ = num_labels UpperCamelCase__ = scope UpperCamelCase__ = n_targets UpperCamelCase__ = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens UpperCamelCase__ = (image_size[1] // patch_size) * (image_size[0] // patch_size) UpperCamelCase__ = num_patches + 1 + self.num_detection_tokens def _lowerCamelCase ( self ): UpperCamelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) UpperCamelCase__ = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) UpperCamelCase__ = [] for i in range(self.batch_size ): UpperCamelCase__ = {} UpperCamelCase__ = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=__lowerCAmelCase ) UpperCamelCase__ = torch.rand(self.n_targets , 4 , device=__lowerCAmelCase ) labels.append(__lowerCAmelCase ) UpperCamelCase__ = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self ): return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = YolosModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = model(__lowerCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = YolosForObjectDetection(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = model(pixel_values=__lowerCAmelCase ) UpperCamelCase__ = model(__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) UpperCamelCase__ = model(pixel_values=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.prepare_config_and_inputs() UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = config_and_inputs UpperCamelCase__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case : Union[str, Any] = (YolosModel, YolosForObjectDetection) if is_torch_available() else () snake_case : Union[str, Any] = ( {"""feature-extraction""": YolosModel, """object-detection""": YolosForObjectDetection} if is_torch_available() else {} ) snake_case : Optional[Any] = False snake_case : List[str] = False snake_case : Dict = False snake_case : List[Any] = False def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ): UpperCamelCase__ = super()._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase , return_labels=__lowerCAmelCase ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": UpperCamelCase__ = [] for i in range(self.model_tester.batch_size ): UpperCamelCase__ = {} UpperCamelCase__ = torch.ones( size=(self.model_tester.n_targets,) , device=__lowerCAmelCase , dtype=torch.long ) UpperCamelCase__ = torch.ones( self.model_tester.n_targets , 4 , device=__lowerCAmelCase , dtype=torch.float ) labels.append(__lowerCAmelCase ) UpperCamelCase__ = labels return inputs_dict def _lowerCamelCase ( self ): UpperCamelCase__ = YolosModelTester(self ) UpperCamelCase__ = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 ) def _lowerCamelCase ( self ): self.config_tester.run_common_tests() def _lowerCamelCase ( self ): # YOLOS does not use inputs_embeds pass def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) UpperCamelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase__ = [*signature.parameters.keys()] UpperCamelCase__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase__ = True # in YOLOS, the seq_len is different UpperCamelCase__ = self.model_tester.expected_seq_len for model_class in self.all_model_classes: UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = True UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCamelCase__ = True UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) UpperCamelCase__ = len(__lowerCAmelCase ) # Check attention is always last and order is fine UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = 1 self.assertEqual(out_len + added_hidden_states , len(__lowerCAmelCase ) ) UpperCamelCase__ = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def _lowerCamelCase ( self ): def check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs.hidden_states UpperCamelCase__ = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__lowerCAmelCase ) , __lowerCAmelCase ) # YOLOS has a different seq_length UpperCamelCase__ = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase__ = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*__lowerCAmelCase ) @slow def _lowerCamelCase ( self ): for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase__ = YolosModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self ): return AutoImageProcessor.from_pretrained("""hustvl/yolos-small""" ) if is_vision_available() else None @slow def _lowerCamelCase ( self ): UpperCamelCase__ = YolosForObjectDetection.from_pretrained("""hustvl/yolos-small""" ).to(__lowerCAmelCase ) UpperCamelCase__ = self.default_image_processor UpperCamelCase__ = prepare_img() UpperCamelCase__ = image_processor(images=__lowerCAmelCase , return_tensors="""pt""" ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCamelCase__ = model(inputs.pixel_values ) # verify outputs UpperCamelCase__ = torch.Size((1, 100, 92) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) UpperCamelCase__ = torch.tensor( [[-24.0248, -10.3024, -14.8290], [-42.0392, -16.8200, -27.4334], [-27.2743, -11.8154, -18.7148]] , device=__lowerCAmelCase , ) UpperCamelCase__ = torch.tensor( [[0.2559, 0.5455, 0.4706], [0.2989, 0.7279, 0.1875], [0.7732, 0.4017, 0.4462]] , device=__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , __lowerCAmelCase , atol=1E-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , __lowerCAmelCase , atol=1E-4 ) ) # verify postprocessing UpperCamelCase__ = image_processor.post_process_object_detection( __lowerCAmelCase , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] UpperCamelCase__ = torch.tensor([0.9994, 0.9790, 0.9964, 0.9972, 0.9861] ).to(__lowerCAmelCase ) UpperCamelCase__ = [75, 75, 17, 63, 17] UpperCamelCase__ = torch.tensor([335.0609, 79.3848, 375.4216, 187.2495] ).to(__lowerCAmelCase ) self.assertEqual(len(results["""scores"""] ) , 5 ) self.assertTrue(torch.allclose(results["""scores"""] , __lowerCAmelCase , atol=1E-4 ) ) self.assertSequenceEqual(results["""labels"""].tolist() , __lowerCAmelCase ) self.assertTrue(torch.allclose(results["""boxes"""][0, :] , __lowerCAmelCase ) )
87
from datetime import datetime as dt import os from github import Github UpperCamelCase__ = [ "good first issue", "good second issue", "good difficult issue", "feature request", "new model", "wip", ] def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = Github(os.environ["""GITHUB_TOKEN"""] ) UpperCamelCase__ = g.get_repo("""huggingface/transformers""" ) UpperCamelCase__ = repo.get_issues(state="""open""" ) for issue in open_issues: UpperCamelCase__ = sorted([comment for comment in issue.get_comments()] , key=lambda a__ : i.created_at , reverse=a__ ) UpperCamelCase__ = comments[0] if len(a__ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) if __name__ == "__main__": main()
87
1
"""simple docstring""" import logging import os from .state import PartialState class UpperCAmelCase_ ( logging.LoggerAdapter): @staticmethod def _UpperCAmelCase ( a ) -> Dict: lowercase__ : Any = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def _UpperCAmelCase ( self , a , a , *a , **a ) -> Union[str, Any]: if PartialState._shared_state == {}: raise RuntimeError( 'You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.' ) lowercase__ : str = kwargs.pop('main_process_only' , a ) lowercase__ : Optional[int] = kwargs.pop('in_order' , a ) if self.isEnabledFor(a ): if self._should_log(a ): lowercase__ , lowercase__ : int = self.process(a , a ) self.logger.log(a , a , *a , **a ) elif in_order: lowercase__ : Dict = PartialState() for i in range(state.num_processes ): if i == state.process_index: lowercase__ , lowercase__ : Optional[Any] = self.process(a , a ) self.logger.log(a , a , *a , **a ) state.wait_for_everyone() def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : str = None ): '''simple docstring''' if log_level is None: lowercase__ : Optional[Any] = os.environ.get('ACCELERATE_LOG_LEVEL' , _lowerCAmelCase ) lowercase__ : List[Any] = logging.getLogger(_lowerCAmelCase ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(_lowerCAmelCase , {} )
77
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _a ( SCREAMING_SNAKE_CASE_ : Optional[int] ): __lowerCAmelCase = filter(lambda SCREAMING_SNAKE_CASE_ : p.requires_grad , model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params UpperCamelCase__ = logging.getLogger(__name__) def _a ( SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Any ): if metric == "rouge2": __lowerCAmelCase = "{val_avg_rouge2:.4f}-{step_count}" elif metric == "bleu": __lowerCAmelCase = "{val_avg_bleu:.4f}-{step_count}" elif metric == "em": __lowerCAmelCase = "{val_avg_em:.4f}-{step_count}" else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" " function." ) __lowerCAmelCase = ModelCheckpoint( dirpath=SCREAMING_SNAKE_CASE_ , filename=SCREAMING_SNAKE_CASE_ , monitor=F"""val_{metric}""" , mode="max" , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def _a ( SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Union[str, Any] ): return EarlyStopping( monitor=F"""val_{metric}""" , mode="min" if "loss" in metric else "max" , patience=SCREAMING_SNAKE_CASE_ , verbose=SCREAMING_SNAKE_CASE_ , ) class a__ ( pl.Callback ): def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" __lowerCAmelCase = {f"""lr_group_{i}""": param["lr"] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(_A ) @rank_zero_only def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A=True ): """simple docstring""" logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / "test_results.txt" __lowerCAmelCase = od / "test_generations.txt" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=_A ) generations_file.parent.mkdir(exist_ok=_A ) with open(_A , "a+" ) as writer: for key in sorted(_A ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(_A , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(_A ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = "\n".join(metrics["preds"] ) generations_file.open("w+" ).write(_A ) @rank_zero_only def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(_A ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1E6, "grad_mp": n_trainable_pars / 1E6} ) @rank_zero_only def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(_A , _A , "test" ) @rank_zero_only def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
92
0
from __future__ import annotations from typing import Generic, TypeVar lowerCAmelCase : Optional[Any] = TypeVar("""T""") class __lowercase ( Generic[T] ): """simple docstring""" def __init__( self : Any , lowerCAmelCase__ : T): SCREAMING_SNAKE_CASE_: Tuple = data SCREAMING_SNAKE_CASE_: List[Any] = self SCREAMING_SNAKE_CASE_: List[str] = 0 class __lowercase ( Generic[T] ): """simple docstring""" def __init__( self : Dict): SCREAMING_SNAKE_CASE_: int = {} def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase__ : T): SCREAMING_SNAKE_CASE_: Optional[Any] = DisjointSetTreeNode(__UpperCAmelCase) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase__ : T): SCREAMING_SNAKE_CASE_: Any = self.map[data] if elem_ref != elem_ref.parent: SCREAMING_SNAKE_CASE_: Optional[Any] = self.find_set(elem_ref.parent.data) return elem_ref.parent def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : DisjointSetTreeNode[T] , lowerCAmelCase__ : DisjointSetTreeNode[T]): if nodea.rank > nodea.rank: SCREAMING_SNAKE_CASE_: Any = nodea else: SCREAMING_SNAKE_CASE_: Dict = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : T , lowerCAmelCase__ : T): self.link(self.find_set(__UpperCAmelCase) , self.find_set(__UpperCAmelCase)) class __lowercase ( Generic[T] ): """simple docstring""" def __init__( self : str): SCREAMING_SNAKE_CASE_: int = {} def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase__ : T): if node not in self.connections: SCREAMING_SNAKE_CASE_: Tuple = {} def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : T , lowerCAmelCase__ : T , lowerCAmelCase__ : int): self.add_node(__UpperCAmelCase) self.add_node(__UpperCAmelCase) SCREAMING_SNAKE_CASE_: List[str] = weight SCREAMING_SNAKE_CASE_: Dict = weight def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: Union[str, Any] = [] SCREAMING_SNAKE_CASE_: str = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start)) edges.append((start, end, self.connections[start][end])) edges.sort(key=lambda lowerCAmelCase__: x[2]) # creating the disjoint set SCREAMING_SNAKE_CASE_: Dict = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(__UpperCAmelCase) # MST generation SCREAMING_SNAKE_CASE_: Union[str, Any] = 0 SCREAMING_SNAKE_CASE_: List[str] = 0 SCREAMING_SNAKE_CASE_: int = GraphUndirectedWeighted[T]() while num_edges < len(self.connections) - 1: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = edges[index] index += 1 SCREAMING_SNAKE_CASE_: Optional[Any] = disjoint_set.find_set(__UpperCAmelCase) SCREAMING_SNAKE_CASE_: str = disjoint_set.find_set(__UpperCAmelCase) if parent_u != parent_v: num_edges += 1 graph.add_edge(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) disjoint_set.union(__UpperCAmelCase , __UpperCAmelCase) return graph
354
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class __lowercase ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[Any]=7 , lowerCAmelCase__ : Tuple=3 , lowerCAmelCase__ : Optional[int]=30 , lowerCAmelCase__ : Dict=400 , lowerCAmelCase__ : int=True , lowerCAmelCase__ : Dict=None , lowerCAmelCase__ : Union[str, Any]=True , lowerCAmelCase__ : Any=[0.5, 0.5, 0.5] , lowerCAmelCase__ : Optional[Any]=[0.5, 0.5, 0.5] , lowerCAmelCase__ : List[Any]=True , lowerCAmelCase__ : Tuple=1 / 255 , lowerCAmelCase__ : int=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p SCREAMING_SNAKE_CASE_: Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} SCREAMING_SNAKE_CASE_: str = parent SCREAMING_SNAKE_CASE_: Tuple = batch_size SCREAMING_SNAKE_CASE_: Tuple = num_channels SCREAMING_SNAKE_CASE_: Union[str, Any] = min_resolution SCREAMING_SNAKE_CASE_: Tuple = max_resolution SCREAMING_SNAKE_CASE_: List[Any] = do_resize SCREAMING_SNAKE_CASE_: Optional[int] = size SCREAMING_SNAKE_CASE_: Optional[int] = do_normalize SCREAMING_SNAKE_CASE_: Any = image_mean SCREAMING_SNAKE_CASE_: Dict = image_std SCREAMING_SNAKE_CASE_: Tuple = do_rescale SCREAMING_SNAKE_CASE_: int = rescale_factor SCREAMING_SNAKE_CASE_: int = do_pad def _SCREAMING_SNAKE_CASE ( self : Optional[int]): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : Dict , lowerCAmelCase__ : int=False): if not batched: SCREAMING_SNAKE_CASE_: List[str] = image_inputs[0] if isinstance(lowerCAmelCase__ , Image.Image): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = image.size else: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = image.shape[1], image.shape[2] if w < h: SCREAMING_SNAKE_CASE_: List[Any] = int(self.size["shortest_edge"] * h / w) SCREAMING_SNAKE_CASE_: Union[str, Any] = self.size["shortest_edge"] elif w > h: SCREAMING_SNAKE_CASE_: Any = self.size["shortest_edge"] SCREAMING_SNAKE_CASE_: Union[str, Any] = int(self.size["shortest_edge"] * w / h) else: SCREAMING_SNAKE_CASE_: int = self.size["shortest_edge"] SCREAMING_SNAKE_CASE_: Dict = self.size["shortest_edge"] else: SCREAMING_SNAKE_CASE_: int = [] for image in image_inputs: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) SCREAMING_SNAKE_CASE_: Tuple = max(lowerCAmelCase__ , key=lambda lowerCAmelCase__: item[0])[0] SCREAMING_SNAKE_CASE_: Optional[Any] = max(lowerCAmelCase__ , key=lambda lowerCAmelCase__: item[1])[1] return expected_height, expected_width @require_torch @require_vision class __lowercase ( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Any = DeformableDetrImageProcessor if is_vision_available() else None def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_: int = DeformableDetrImageProcessingTester(self) @property def _SCREAMING_SNAKE_CASE ( self : Optional[int]): return self.image_processor_tester.prepare_image_processor_dict() def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_: Dict = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(lowerCAmelCase__ , "image_mean")) self.assertTrue(hasattr(lowerCAmelCase__ , "image_std")) self.assertTrue(hasattr(lowerCAmelCase__ , "do_normalize")) self.assertTrue(hasattr(lowerCAmelCase__ , "do_resize")) self.assertTrue(hasattr(lowerCAmelCase__ , "do_rescale")) self.assertTrue(hasattr(lowerCAmelCase__ , "do_pad")) self.assertTrue(hasattr(lowerCAmelCase__ , "size")) def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Dict = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333}) self.assertEqual(image_processor.do_pad , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCAmelCase__) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84}) self.assertEqual(image_processor.do_pad , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): pass def _SCREAMING_SNAKE_CASE ( self : List[Any]): # Initialize image_processing SCREAMING_SNAKE_CASE_: List[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images SCREAMING_SNAKE_CASE_: Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image) # Test not batched input SCREAMING_SNAKE_CASE_: Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt").pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[str] = self.image_processor_tester.get_expected_values(lowerCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[Any] = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = image_processing(lowerCAmelCase__ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _SCREAMING_SNAKE_CASE ( self : str): # Initialize image_processing SCREAMING_SNAKE_CASE_: Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors SCREAMING_SNAKE_CASE_: List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray) # Test not batched input SCREAMING_SNAKE_CASE_: str = image_processing(image_inputs[0] , return_tensors="pt").pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[str] = self.image_processor_tester.get_expected_values(lowerCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched SCREAMING_SNAKE_CASE_: Any = image_processing(lowerCAmelCase__ , return_tensors="pt").pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: int = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _SCREAMING_SNAKE_CASE ( self : List[Any]): # Initialize image_processing SCREAMING_SNAKE_CASE_: List[Any] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors SCREAMING_SNAKE_CASE_: int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor) # Test not batched input SCREAMING_SNAKE_CASE_: Dict = image_processing(image_inputs[0] , return_tensors="pt").pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = self.image_processor_tester.get_expected_values(lowerCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched SCREAMING_SNAKE_CASE_: Optional[Any] = image_processing(lowerCAmelCase__ , return_tensors="pt").pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[Any] = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _SCREAMING_SNAKE_CASE ( self : Tuple): # prepare image and target SCREAMING_SNAKE_CASE_: Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r") as f: SCREAMING_SNAKE_CASE_: str = json.loads(f.read()) SCREAMING_SNAKE_CASE_: Optional[int] = {"image_id": 3_9769, "annotations": target} # encode them SCREAMING_SNAKE_CASE_: str = DeformableDetrImageProcessor() SCREAMING_SNAKE_CASE_: Dict = image_processing(images=lowerCAmelCase__ , annotations=lowerCAmelCase__ , return_tensors="pt") # verify pixel values SCREAMING_SNAKE_CASE_: Optional[Any] = torch.Size([1, 3, 800, 1066]) self.assertEqual(encoding["pixel_values"].shape , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCAmelCase__ , atol=1E-4)) # verify area SCREAMING_SNAKE_CASE_: int = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438]) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCAmelCase__)) # verify boxes SCREAMING_SNAKE_CASE_: str = torch.Size([6, 4]) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215]) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCAmelCase__ , atol=1E-3)) # verify image_id SCREAMING_SNAKE_CASE_: str = torch.tensor([3_9769]) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCAmelCase__)) # verify is_crowd SCREAMING_SNAKE_CASE_: int = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCAmelCase__)) # verify class_labels SCREAMING_SNAKE_CASE_: Tuple = torch.tensor([75, 75, 63, 65, 17, 17]) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCAmelCase__)) # verify orig_size SCREAMING_SNAKE_CASE_: Union[str, Any] = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCAmelCase__)) # verify size SCREAMING_SNAKE_CASE_: str = torch.tensor([800, 1066]) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCAmelCase__)) @slow def _SCREAMING_SNAKE_CASE ( self : Tuple): # prepare image, target and masks_path SCREAMING_SNAKE_CASE_: Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r") as f: SCREAMING_SNAKE_CASE_: List[Any] = json.loads(f.read()) SCREAMING_SNAKE_CASE_: Optional[Any] = {"file_name": "000000039769.png", "image_id": 3_9769, "segments_info": target} SCREAMING_SNAKE_CASE_: int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic") # encode them SCREAMING_SNAKE_CASE_: Any = DeformableDetrImageProcessor(format="coco_panoptic") SCREAMING_SNAKE_CASE_: Optional[Any] = image_processing(images=lowerCAmelCase__ , annotations=lowerCAmelCase__ , masks_path=lowerCAmelCase__ , return_tensors="pt") # verify pixel values SCREAMING_SNAKE_CASE_: Dict = torch.Size([1, 3, 800, 1066]) self.assertEqual(encoding["pixel_values"].shape , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCAmelCase__ , atol=1E-4)) # verify area SCREAMING_SNAKE_CASE_: Optional[Any] = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147]) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCAmelCase__)) # verify boxes SCREAMING_SNAKE_CASE_: List[str] = torch.Size([6, 4]) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625]) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCAmelCase__ , atol=1E-3)) # verify image_id SCREAMING_SNAKE_CASE_: Any = torch.tensor([3_9769]) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCAmelCase__)) # verify is_crowd SCREAMING_SNAKE_CASE_: Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCAmelCase__)) # verify class_labels SCREAMING_SNAKE_CASE_: List[Any] = torch.tensor([17, 17, 63, 75, 75, 93]) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCAmelCase__)) # verify masks SCREAMING_SNAKE_CASE_: Tuple = 82_2873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCAmelCase__) # verify orig_size SCREAMING_SNAKE_CASE_: str = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCAmelCase__)) # verify size SCREAMING_SNAKE_CASE_: Optional[int] = torch.tensor([800, 1066]) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCAmelCase__))
127
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable def __UpperCamelCase ( lowercase__ : Tuple, lowercase__ : List[str], lowercase__ : int, lowercase__ : str = 1_00, ): '''simple docstring''' __lowercase =x_start __lowercase =fnc(_UpperCamelCase ) __lowercase =0.0 for _ in range(_UpperCamelCase ): # Approximates small segments of curve as linear and solve # for trapezoidal area __lowercase =(x_end - x_start) / steps + xa __lowercase =fnc(_UpperCamelCase ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step __lowercase =xa __lowercase =fxa return area if __name__ == "__main__": def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') UpperCAmelCase = 10 while i <= 10_0000: print(F'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
141
from __future__ import annotations def __lowercase ( _UpperCamelCase ) ->float: """simple docstring""" if not nums: raise ValueError('''List is empty''' ) return sum(_UpperCamelCase ) / len(_UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
337
0
'''simple docstring''' import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("9.1.0"): UpperCamelCase_ = { "linear": PIL.Image.Resampling.BILINEAR, "bilinear": PIL.Image.Resampling.BILINEAR, "bicubic": PIL.Image.Resampling.BICUBIC, "lanczos": PIL.Image.Resampling.LANCZOS, "nearest": PIL.Image.Resampling.NEAREST, } else: UpperCamelCase_ = { "linear": PIL.Image.LINEAR, "bilinear": PIL.Image.BILINEAR, "bicubic": PIL.Image.BICUBIC, "lanczos": PIL.Image.LANCZOS, "nearest": PIL.Image.NEAREST, } def lowercase__( __UpperCamelCase: int ): """simple docstring""" SCREAMING_SNAKE_CASE : int = (images / 2 + 0.5).clamp(0 ,1 ) SCREAMING_SNAKE_CASE : List[str] = images.cpu().permute(0 ,2 ,3 ,1 ).float().numpy() SCREAMING_SNAKE_CASE : Any = numpy_to_pil(__UpperCamelCase ) return images def lowercase__( __UpperCamelCase: Optional[int] ): """simple docstring""" if images.ndim == 3: SCREAMING_SNAKE_CASE : Dict = images[None, ...] SCREAMING_SNAKE_CASE : Optional[Any] = (images * 2_55).round().astype('uint8' ) if images.shape[-1] == 1: # special case for grayscale (single channel) images SCREAMING_SNAKE_CASE : List[Any] = [Image.fromarray(image.squeeze() ,mode='L' ) for image in images] else: SCREAMING_SNAKE_CASE : Tuple = [Image.fromarray(__UpperCamelCase ) for image in images] return pil_images
246
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase_ = { "configuration_mobilebert": [ "MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MobileBertConfig", "MobileBertOnnxConfig", ], "tokenization_mobilebert": ["MobileBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = ["MobileBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ "MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "MobileBertForMaskedLM", "MobileBertForMultipleChoice", "MobileBertForNextSentencePrediction", "MobileBertForPreTraining", "MobileBertForQuestionAnswering", "MobileBertForSequenceClassification", "MobileBertForTokenClassification", "MobileBertLayer", "MobileBertModel", "MobileBertPreTrainedModel", "load_tf_weights_in_mobilebert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ "TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFMobileBertForMaskedLM", "TFMobileBertForMultipleChoice", "TFMobileBertForNextSentencePrediction", "TFMobileBertForPreTraining", "TFMobileBertForQuestionAnswering", "TFMobileBertForSequenceClassification", "TFMobileBertForTokenClassification", "TFMobileBertMainLayer", "TFMobileBertModel", "TFMobileBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
246
1
'''simple docstring''' def _A ( A__ ): """simple docstring""" return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
104
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class UpperCAmelCase__ ( A__ ): """simple docstring""" a = (UnCLIPScheduler,) def lowercase_ ( self : List[str] , **__lowerCamelCase : int ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = { '''num_train_timesteps''': 1000, '''variance_type''': '''fixed_small_log''', '''clip_sample''': True, '''clip_sample_range''': 1.0, '''prediction_type''': '''epsilon''', } config.update(**__lowerCamelCase ) return config def lowercase_ ( self : Dict ) -> Any: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=__lowerCamelCase ) def lowercase_ ( self : str ) -> Union[str, Any]: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=__lowerCamelCase ) def lowercase_ ( self : List[str] ) -> int: for clip_sample in [True, False]: self.check_over_configs(clip_sample=__lowerCamelCase ) def lowercase_ ( self : Optional[Any] ) -> Tuple: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=__lowerCamelCase ) def lowercase_ ( self : Union[str, Any] ) -> Dict: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=__lowerCamelCase ) def lowercase_ ( self : int ) -> str: for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=__lowerCamelCase , prev_timestep=__lowerCamelCase ) def lowercase_ ( self : Dict ) -> Dict: SCREAMING_SNAKE_CASE__ = self.scheduler_classes[0] SCREAMING_SNAKE_CASE__ = self.get_scheduler_config(variance_type='''fixed_small_log''' ) SCREAMING_SNAKE_CASE__ = scheduler_class(**__lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.00_00e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0549625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.9994987 ) ) < 1e-5 def lowercase_ ( self : Union[str, Any] ) -> int: SCREAMING_SNAKE_CASE__ = self.scheduler_classes[0] SCREAMING_SNAKE_CASE__ = self.get_scheduler_config(variance_type='''learned_range''' ) SCREAMING_SNAKE_CASE__ = scheduler_class(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = 0.5 assert scheduler._get_variance(1 , predicted_variance=__lowerCamelCase ) - -10.1712790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=__lowerCamelCase ) - -5.7998052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=__lowerCamelCase ) - -0.0010011 < 1e-5 def lowercase_ ( self : Tuple ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = self.scheduler_classes[0] SCREAMING_SNAKE_CASE__ = self.get_scheduler_config() SCREAMING_SNAKE_CASE__ = scheduler_class(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = scheduler.timesteps SCREAMING_SNAKE_CASE__ = self.dummy_model() SCREAMING_SNAKE_CASE__ = self.dummy_sample_deter SCREAMING_SNAKE_CASE__ = torch.manual_seed(0 ) for i, t in enumerate(__lowerCamelCase ): # 1. predict noise residual SCREAMING_SNAKE_CASE__ = model(__lowerCamelCase , __lowerCamelCase ) # 2. predict previous mean of sample x_t-1 SCREAMING_SNAKE_CASE__ = scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase ).prev_sample SCREAMING_SNAKE_CASE__ = pred_prev_sample SCREAMING_SNAKE_CASE__ = torch.sum(torch.abs(__lowerCamelCase ) ) SCREAMING_SNAKE_CASE__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 252.2682495 ) < 1e-2 assert abs(result_mean.item() - 0.3284743 ) < 1e-3 def lowercase_ ( self : Tuple ) -> Dict: SCREAMING_SNAKE_CASE__ = self.scheduler_classes[0] SCREAMING_SNAKE_CASE__ = self.get_scheduler_config() SCREAMING_SNAKE_CASE__ = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(25 ) SCREAMING_SNAKE_CASE__ = scheduler.timesteps SCREAMING_SNAKE_CASE__ = self.dummy_model() SCREAMING_SNAKE_CASE__ = self.dummy_sample_deter SCREAMING_SNAKE_CASE__ = torch.manual_seed(0 ) for i, t in enumerate(__lowerCamelCase ): # 1. predict noise residual SCREAMING_SNAKE_CASE__ = model(__lowerCamelCase , __lowerCamelCase ) if i + 1 == timesteps.shape[0]: SCREAMING_SNAKE_CASE__ = None else: SCREAMING_SNAKE_CASE__ = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 SCREAMING_SNAKE_CASE__ = scheduler.step( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , prev_timestep=__lowerCamelCase , generator=__lowerCamelCase ).prev_sample SCREAMING_SNAKE_CASE__ = pred_prev_sample SCREAMING_SNAKE_CASE__ = torch.sum(torch.abs(__lowerCamelCase ) ) SCREAMING_SNAKE_CASE__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 258.2044983 ) < 1e-2 assert abs(result_mean.item() - 0.3362038 ) < 1e-3 def lowercase_ ( self : int ) -> Tuple: pass def lowercase_ ( self : Dict ) -> Union[str, Any]: pass
314
0
import os def lowerCamelCase_ ( ): with open(os.path.dirname(_lowerCamelCase ) + '/p022_names.txt' ) as file: lowerCamelCase__ : str = str(file.readlines()[0] ) lowerCamelCase__ : Optional[int] = names.replace('\"' , '' ).split(',' ) names.sort() lowerCamelCase__ : Any = 0 lowerCamelCase__ : List[Any] = 0 for i, name in enumerate(_lowerCamelCase ): for letter in name: name_score += ord(_lowerCamelCase ) - 64 total_score += (i + 1) * name_score lowerCamelCase__ : str = 0 return total_score if __name__ == "__main__": print(solution())
360
"""simple docstring""" def lowerCamelCase_ ( _lowerCamelCase ): lowerCamelCase__ : Union[str, Any] = [] lowerCamelCase__ : List[str] = [] lowerCamelCase__ : Tuple = { '^': 3, '*': 2, '/': 2, '%': 2, '+': 1, '-': 1, } # Priority of each operator lowerCamelCase__ : List[str] = len(_lowerCamelCase ) if (len(_lowerCamelCase ) > 7) else 7 # Print table header for output print( 'Symbol'.center(8 ) , 'Stack'.center(_lowerCamelCase ) , 'Postfix'.center(_lowerCamelCase ) , sep=' | ' , ) print('-' * (print_width * 3 + 7) ) for x in infix: if x.isalpha() or x.isdigit(): post_fix.append(_lowerCamelCase ) # if x is Alphabet / Digit, add it to Postfix elif x == "(": stack.append(_lowerCamelCase ) # if x is "(" push to Stack elif x == ")": # if x is ")" pop stack until "(" is encountered while stack[-1] != "(": post_fix.append(stack.pop() ) # Pop stack & add the content to Postfix stack.pop() else: if len(_lowerCamelCase ) == 0: stack.append(_lowerCamelCase ) # If stack is empty, push x to stack else: # while priority of x is not > priority of element in the stack while len(_lowerCamelCase ) > 0 and priority[x] <= priority[stack[-1]]: post_fix.append(stack.pop() ) # pop stack & add to Postfix stack.append(_lowerCamelCase ) # push x to stack print( x.center(8 ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , sep=' | ' , ) # Output in tabular format while len(_lowerCamelCase ) > 0: # while stack is not empty post_fix.append(stack.pop() ) # pop stack & add to Postfix print( ' '.center(8 ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , sep=' | ' , ) # Output in tabular format return "".join(_lowerCamelCase ) # return Postfix as str def lowerCamelCase_ ( _lowerCamelCase ): lowerCamelCase__ : Union[str, Any] = list(infix[::-1] ) # reverse the infix equation for i in range(len(_lowerCamelCase ) ): if infix[i] == "(": lowerCamelCase__ : List[Any] = ')' # change "(" to ")" elif infix[i] == ")": lowerCamelCase__ : Tuple = '(' # change ")" to "(" return (infix_2_postfix(''.join(_lowerCamelCase ) ))[ ::-1 ] # call infix_2_postfix on Infix, return reverse of Postfix if __name__ == "__main__": A_ : Tuple = input("\nEnter an Infix Equation = ") # Input an Infix equation A_ : List[str] = "".join(Infix.split()) # Remove spaces from the input print("\n\t", Infix, "(Infix) -> ", infix_2_prefix(Infix), "(Prefix)")
316
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: if not nums: # Makes sure that the list is not empty raise ValueError('List is empty' ) lowerCAmelCase__ : str = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
212
from math import pi def snake_case_ ( lowerCAmelCase_ : int , lowerCAmelCase_ : int ): return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
233
0
"""simple docstring""" import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Any , _UpperCAmelCase : Tuple , _UpperCAmelCase : int ): # Initialise PyTorch model lowerCAmelCase = TaConfig.from_json_file(_UpperCAmelCase ) print(F'Building PyTorch model from configuration: {config}' ) lowerCAmelCase = TaForConditionalGeneration(_UpperCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_ta(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __UpperCamelCase : Union[str, Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
363
"""simple docstring""" from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __UpperCamelCase : str = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __UpperCamelCase : Optional[Any] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __UpperCamelCase : Dict = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : str ): lowerCAmelCase = len([g for position, g in enumerate(_UpperCAmelCase ) if g == main_target[position]] ) return (item, float(_UpperCAmelCase )) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : str ): lowerCAmelCase = random.randint(0 , len(_UpperCAmelCase ) - 1 ) lowerCAmelCase = parent_a[:random_slice] + parent_a[random_slice:] lowerCAmelCase = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : list[str] ): lowerCAmelCase = list(_UpperCAmelCase ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: lowerCAmelCase = random.choice(_UpperCAmelCase ) return "".join(_UpperCAmelCase ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : tuple[str, float] , _UpperCAmelCase : list[tuple[str, float]] , _UpperCAmelCase : list[str] , ): lowerCAmelCase = [] # Generate more children proportionally to the fitness score. lowerCAmelCase = int(parent_a[1] * 100 ) + 1 lowerCAmelCase = 10 if child_n >= 10 else child_n for _ in range(_UpperCAmelCase ): lowerCAmelCase = population_score[random.randint(0 , _UpperCAmelCase )][0] lowerCAmelCase ,lowerCAmelCase = crossover(parent_a[0] , _UpperCAmelCase ) # Append new string to the population list. pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) return pop def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : list[str] , _UpperCAmelCase : bool = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: lowerCAmelCase = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(_UpperCAmelCase ) # Verify that the target contains no genes besides the ones inside genes variable. lowerCAmelCase = sorted({c for c in target if c not in genes} ) if not_in_genes_list: lowerCAmelCase = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(_UpperCAmelCase ) # Generate random starting population. lowerCAmelCase = [] for _ in range(_UpperCAmelCase ): population.append(''.join([random.choice(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) )] ) ) # Just some logs to know what the algorithms is doing. lowerCAmelCase ,lowerCAmelCase = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(_UpperCAmelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. lowerCAmelCase = [evaluate(_UpperCAmelCase , _UpperCAmelCase ) for item in population] # Check if there is a matching evolution. lowerCAmelCase = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x[1] , reverse=_UpperCAmelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. lowerCAmelCase = population[: int(N_POPULATION / 3 )] population.clear() population.extend(_UpperCAmelCase ) # Normalize population score to be between 0 and 1. lowerCAmelCase = [ (item, score / len(_UpperCAmelCase )) for item, score in population_score ] # This is selection for i in range(_UpperCAmelCase ): population.extend(select(population_score[int(_UpperCAmelCase )] , _UpperCAmelCase , _UpperCAmelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(_UpperCAmelCase ) > N_POPULATION: break if __name__ == "__main__": __UpperCamelCase : Tuple = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __UpperCamelCase : str = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase : Dict = basic(target_str, genes_list) print( f'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
309
0
"""simple docstring""" import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def __UpperCAmelCase ( lowercase ): """simple docstring""" return 1.0 / (1.0 + np.exp(-_outputs )) def __UpperCAmelCase ( lowercase ): """simple docstring""" _UpperCAmelCase = np.max(_outputs ,axis=-1 ,keepdims=SCREAMING_SNAKE_CASE_ ) _UpperCAmelCase = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 ,keepdims=SCREAMING_SNAKE_CASE_ ) class a ( __snake_case ): _snake_case : Optional[int] = """sigmoid""" _snake_case : List[Any] = """softmax""" _snake_case : Optional[Any] = """none""" @add_end_docstrings( __snake_case , R'\n return_all_scores (`bool`, *optional*, defaults to `False`):\n Whether to return all prediction scores or just the one of the predicted class.\n function_to_apply (`str`, *optional*, defaults to `\"default\"`):\n The function to apply to the model outputs in order to retrieve the scores. Accepts four different values:\n\n - `\"default\"`: if the model has a single label, will apply the sigmoid function on the output. If the model\n has several labels, will apply the softmax function on the output.\n - `\"sigmoid\"`: Applies the sigmoid function on the output.\n - `\"softmax\"`: Applies the softmax function on the output.\n - `\"none\"`: Does not apply any function on the output.\n ' , ) class a ( __snake_case ): _snake_case : Dict = False _snake_case : List[str] = ClassificationFunction.NONE def __init__( self : Optional[Any] , **__lowerCAmelCase : List[Any] ): super().__init__(**__lowerCAmelCase ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def lowerCAmelCase_ ( self : Dict , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Union[str, Any]=None , __lowerCAmelCase : List[str]="" , **__lowerCAmelCase : int ): # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" _UpperCAmelCase = tokenizer_kwargs _UpperCAmelCase = {} if hasattr(self.model.config , """return_all_scores""" ) and return_all_scores is None: _UpperCAmelCase = self.model.config.return_all_scores if isinstance(__lowerCAmelCase , __lowerCAmelCase ) or top_k is None: _UpperCAmelCase = top_k _UpperCAmelCase = False elif return_all_scores is not None: warnings.warn( """`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of""" """ `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.""" , __lowerCAmelCase , ) if return_all_scores: _UpperCAmelCase = None else: _UpperCAmelCase = 1 if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: _UpperCAmelCase = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self : str , *__lowerCAmelCase : Optional[Any] , **__lowerCAmelCase : Tuple ): _UpperCAmelCase = super().__call__(*__lowerCAmelCase , **__lowerCAmelCase ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. _UpperCAmelCase = 'top_k' not in kwargs if isinstance(args[0] , __lowerCAmelCase ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def lowerCAmelCase_ ( self : int , __lowerCAmelCase : Optional[int] , **__lowerCAmelCase : Union[str, Any] ): _UpperCAmelCase = self.framework if isinstance(__lowerCAmelCase , __lowerCAmelCase ): return self.tokenizer(**__lowerCAmelCase , return_tensors=__lowerCAmelCase , **__lowerCAmelCase ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ) and len(__lowerCAmelCase ) == 1 and isinstance(inputs[0] , __lowerCAmelCase ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=__lowerCAmelCase , **__lowerCAmelCase ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( """The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a""" """ dictionary `{\"text\": \"My text\", \"text_pair\": \"My pair\"}` in order to send a text pair.""" ) return self.tokenizer(__lowerCAmelCase , return_tensors=__lowerCAmelCase , **__lowerCAmelCase ) def lowerCAmelCase_ ( self : Dict , __lowerCAmelCase : Dict ): return self.model(**__lowerCAmelCase ) def lowerCAmelCase_ ( self : str , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : str=1 , __lowerCAmelCase : Optional[int]=True ): # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: _UpperCAmelCase = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: _UpperCAmelCase = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , """function_to_apply""" ) and function_to_apply is None: _UpperCAmelCase = self.model.config.function_to_apply else: _UpperCAmelCase = ClassificationFunction.NONE _UpperCAmelCase = model_outputs['logits'][0] _UpperCAmelCase = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: _UpperCAmelCase = sigmoid(__lowerCAmelCase ) elif function_to_apply == ClassificationFunction.SOFTMAX: _UpperCAmelCase = softmax(__lowerCAmelCase ) elif function_to_apply == ClassificationFunction.NONE: _UpperCAmelCase = outputs else: raise ValueError(f'''Unrecognized `function_to_apply` argument: {function_to_apply}''' ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} _UpperCAmelCase = [ {'label': self.model.config.idalabel[i], 'score': score.item()} for i, score in enumerate(__lowerCAmelCase ) ] if not _legacy: dict_scores.sort(key=lambda __lowerCAmelCase : x["score"] , reverse=__lowerCAmelCase ) if top_k is not None: _UpperCAmelCase = dict_scores[:top_k] return dict_scores
289
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor snake_case_ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ (__snake_case ): def __init__( self , *a , **a): warnings.warn( 'The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use FlavaImageProcessor instead.' , a , ) super().__init__(*a , **a)
214
0
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def snake_case_ ( )-> Tuple: '''simple docstring''' _UpperCAmelCase : Dict = { """repo_name""": ["""test_repo1""", """test_repo2""", """test_repo3"""], """path""": ["""test_1.py""", """test_2.py""", """unit_test.py"""], """content""": ["""a """ * 20, """a """ * 30, """b """ * 7], } _UpperCAmelCase : int = Dataset.from_dict(lowerCAmelCase_ ) return dataset class lowercase ( _lowerCamelCase ): """simple docstring""" def _snake_case ( self ) -> List[Any]: _UpperCAmelCase : Dict = get_dataset() _UpperCAmelCase : Tuple = make_duplicate_clusters(a_ ,0.85 ) self.assertEqual(len(duplicate_clusters[0] ) ,2 ) def _snake_case ( self ) -> Optional[Any]: _UpperCAmelCase : Union[str, Any] = get_dataset() _UpperCAmelCase ,_UpperCAmelCase : Union[str, Any] = deduplicate_dataset(a_ ) self.assertEqual(len(a_ ) ,2 ) print(a_ ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] ,2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] ,a_ )
349
'''simple docstring''' def snake_case_ ( lowerCAmelCase_ )-> int: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""only integers accepted as input""" ) else: _UpperCAmelCase : Dict = str(abs(lowerCAmelCase_ ) ) _UpperCAmelCase : Optional[Any] = [list(lowerCAmelCase_ ) for char in range(len(lowerCAmelCase_ ) )] for index in range(len(lowerCAmelCase_ ) ): num_transpositions[index].pop(lowerCAmelCase_ ) return max( int("""""".join(list(lowerCAmelCase_ ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__("""doctest""").testmod()
349
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase_ = { '''configuration_whisper''': ['''WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WhisperConfig''', '''WhisperOnnxConfig'''], '''feature_extraction_whisper''': ['''WhisperFeatureExtractor'''], '''processing_whisper''': ['''WhisperProcessor'''], '''tokenization_whisper''': ['''WhisperTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['''WhisperTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WhisperForConditionalGeneration''', '''WhisperModel''', '''WhisperPreTrainedModel''', '''WhisperForAudioClassification''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWhisperForConditionalGeneration''', '''TFWhisperModel''', '''TFWhisperPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''FlaxWhisperForConditionalGeneration''', '''FlaxWhisperModel''', '''FlaxWhisperPreTrainedModel''', '''FlaxWhisperForAudioClassification''', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
8
from datetime import datetime as dt import os from github import Github A__ : List[str] = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''feature request''', '''new model''', '''wip''', ] def UpperCamelCase( ): lowerCAmelCase_ : Union[str, Any] = Github(os.environ['''GITHUB_TOKEN'''] ) lowerCAmelCase_ : Tuple = g.get_repo('''huggingface/transformers''' ) lowerCAmelCase_ : int = repo.get_issues(state='''open''' ) for issue in open_issues: lowerCAmelCase_ : Optional[Any] = sorted([comment for comment in issue.get_comments()] ,key=lambda __UpperCamelCase : i.created_at ,reverse=__UpperCamelCase ) lowerCAmelCase_ : Tuple = comments[0] if len(__UpperCamelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state='''closed''' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) if __name__ == "__main__": main()
103
0
import os import re import shutil import sys import tempfile import unittest import black _UpperCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. _UpperCAmelCase = """ def __init__(self, config): super().__init__() self.transform = BertPredictionHeadTransform(config) # The output weights are the same as the input embeddings, but there is # an output-only bias for each token. self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False) self.bias = nn.Parameter(torch.zeros(config.vocab_size)) # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings` self.decoder.bias = self.bias def forward(self, hidden_states): hidden_states = self.transform(hidden_states) hidden_states = self.decoder(hidden_states) return hidden_states """ class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ ( self ): """simple docstring""" A_ : List[str] = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , 'models/bert/' ) ) A_ : Tuple = self.transformer_dir shutil.copy( os.path.join(lowercase , 'src/transformers/models/bert/modeling_bert.py' ) , os.path.join(self.transformer_dir , 'models/bert/modeling_bert.py' ) , ) def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Optional[int] = 'src/transformers' shutil.rmtree(self.transformer_dir ) def lowerCAmelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None ): """simple docstring""" A_ : Optional[Any] = comment + F'''\nclass {class_name}(nn.Module):\n''' + class_code if overwrite_result is not None: A_ : Dict = comment + F'''\nclass {class_name}(nn.Module):\n''' + overwrite_result A_ : Optional[int] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_1_9 ) A_ : int = black.format_str(lowercase , mode=lowercase ) A_ : List[Any] = os.path.join(self.transformer_dir , 'new_code.py' ) with open(lowercase , 'w' , newline='\n' ) as f: f.write(lowercase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(lowercase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=lowercase ) with open(lowercase , 'r' ) as f: self.assertTrue(f.read() , lowercase ) def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Dict = check_copies.find_code_in_transformers('models.bert.modeling_bert.BertLMPredictionHead' ) self.assertEqual(lowercase , lowercase ) def lowerCAmelCase_ ( self ): """simple docstring""" self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead' , 'BertLMPredictionHead' , REFERENCE_CODE + '\n' , ) # With no empty line at the end self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead' , 'BertLMPredictionHead' , lowercase , ) # Copy consistency with rename self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel' , 'TestModelLMPredictionHead' , re.sub('Bert' , 'TestModel' , lowercase ) , ) # Copy consistency with a really long name A_ : Optional[Any] = 'TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason' self.check_copy_consistency( F'''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}''' , F'''{long_class_name}LMPredictionHead''' , re.sub('Bert' , lowercase , lowercase ) , ) # Copy consistency with overwrite self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel' , 'TestModelLMPredictionHead' , lowercase , overwrite_result=re.sub('Bert' , 'TestModel' , lowercase ) , ) def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Tuple = check_copies.LOCALIZED_READMES['README_zh-hans.md'] A_ : str = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the' ' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for' ' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong' ' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.' ' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),' ' released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and' ' lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same' ' method has been applied to compress GPT2 into' ' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into' ' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),' ' Multilingual BERT into' ' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German' ' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**' ' (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders' ' as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang' ' Luong, Quoc V. Le, Christopher D. Manning.' ) A_ : Any = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) A_ : List[str] = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.' ' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文' ' [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and' ' lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same' ' method has been applied to compress GPT2 into' ' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into' ' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),' ' Multilingual BERT into' ' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German' ' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自' ' Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather' ' than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,' ' Christopher D. Manning 发布。\n' ) A_ , A_ : str = check_copies.convert_to_localized_md( lowercase , lowercase , localized_readme['format_model_list'] ) self.assertFalse(lowercase ) self.assertEqual(lowercase , lowercase ) A_ , A_ : Tuple = check_copies.convert_to_localized_md( lowercase , lowercase , localized_readme['format_model_list'] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(lowercase ) A_ : Optional[int] = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the' ' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for' ' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong' ' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.' ) A_ : Tuple = ( '1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and' ' the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) A_ : Dict = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) A_ , A_ : int = check_copies.convert_to_localized_md( lowercase , lowercase , localized_readme['format_model_list'] ) # Check if the model link is synchronized. self.assertEqual(lowercase , lowercase )
192
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict _UpperCAmelCase = namedtuple( """_TestCommandArgs""", [ """dataset""", """name""", """cache_dir""", """data_dir""", """all_configs""", """save_infos""", """ignore_verifications""", """force_redownload""", """clear_cache""", ], defaults=[None, None, None, False, False, False, False, False], ) def UpperCamelCase ( __lowercase : Any ,__lowercase : str ): '''simple docstring''' return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def UpperCamelCase ( __lowercase : Dict ): '''simple docstring''' A_ : Optional[Any] = _TestCommandArgs(dataset=__lowercase ,all_configs=__lowercase ,save_infos=__lowercase ) A_ : List[Any] = TestCommand(*__lowercase ) test_command.run() A_ : Any = os.path.join(__lowercase ,'README.md' ) assert os.path.exists(__lowercase ) A_ : Tuple = DatasetInfosDict.from_directory(__lowercase ) A_ : Any = DatasetInfosDict( { 'default': DatasetInfo( features=Features( { 'tokens': Sequence(Value('string' ) ), 'ner_tags': Sequence( ClassLabel(names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'] ) ), 'langs': Sequence(Value('string' ) ), 'spans': Sequence(Value('string' ) ), } ) ,splits=[ { 'name': 'train', 'num_bytes': 2_35_15_63, 'num_examples': 1_00_00, }, { 'name': 'validation', 'num_bytes': 23_84_18, 'num_examples': 10_00, }, ] ,download_size=3_94_06_80 ,dataset_size=2_58_99_81 ,) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: A_ , A_ : Union[str, Any] = getattr(dataset_infos['default'] ,__lowercase ), getattr(expected_dataset_infos['default'] ,__lowercase ) if key == "num_bytes": assert is_apercent_close(__lowercase ,__lowercase ) elif key == "splits": assert list(__lowercase ) == list(__lowercase ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes ,expected[split].num_bytes ) else: result == expected
192
1
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def lowerCamelCase__ ( a__ : List[str] ) -> str: UpperCamelCase_ = [] for line in lines: UpperCamelCase_ = re.sub(r"""#.*""" , """""" , a__ ) # remove comments if line: filtered_lines.append(a__ ) UpperCamelCase_ = """\n""".join(a__ ) # Make a hash from all this code UpperCamelCase_ = full_str.encode("""utf-8""" ) return shaaaa(a__ ).hexdigest() # get importable module names and hash for caching _A = { '''csv''': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), '''json''': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), '''pandas''': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), '''parquet''': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), '''arrow''': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), '''text''': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), '''imagefolder''': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), '''audiofolder''': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _A = { '''.csv''': ('''csv''', {}), '''.tsv''': ('''csv''', {'''sep''': '''\t'''}), '''.json''': ('''json''', {}), '''.jsonl''': ('''json''', {}), '''.parquet''': ('''parquet''', {}), '''.arrow''': ('''arrow''', {}), '''.txt''': ('''text''', {}), } _EXTENSION_TO_MODULE.update({ext: ('''imagefolder''', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('''imagefolder''', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('''audiofolder''', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('''audiofolder''', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _A = {'''imagefolder''', '''audiofolder'''} # Used to filter data files based on extensions given a module name _A = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('''.zip''') _MODULE_TO_EXTENSIONS["audiofolder"].append('''.zip''')
122
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging _A = logging.get_logger(__name__) if is_vision_available(): import PIL class lowercase_ ( __SCREAMING_SNAKE_CASE ): A__ : Optional[int] = ["""pixel_values"""] def __init__( self , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = PILImageResampling.BICUBIC , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = True , __UpperCamelCase = 1 / 2_5_5 , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = True , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) UpperCamelCase_ = size if size is not None else {"""shortest_edge""": 2_2_4} UpperCamelCase_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) UpperCamelCase_ = crop_size if crop_size is not None else {"""height""": 2_2_4, """width""": 2_2_4} UpperCamelCase_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase , param_name="""crop_size""" ) UpperCamelCase_ = do_resize UpperCamelCase_ = size UpperCamelCase_ = resample UpperCamelCase_ = do_center_crop UpperCamelCase_ = crop_size UpperCamelCase_ = do_rescale UpperCamelCase_ = rescale_factor UpperCamelCase_ = do_normalize UpperCamelCase_ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCamelCase_ = image_std if image_std is not None else OPENAI_CLIP_STD UpperCamelCase_ = do_convert_rgb def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = PILImageResampling.BICUBIC , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" UpperCamelCase_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) UpperCamelCase_ = get_resize_output_image_size(__UpperCamelCase , size=size["""shortest_edge"""] , default_to_square=__UpperCamelCase ) return resize(__UpperCamelCase , size=__UpperCamelCase , resample=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" UpperCamelCase_ = get_size_dict(__UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(__UpperCamelCase , size=(size["""height"""], size["""width"""]) , data_format=__UpperCamelCase , **__UpperCamelCase ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" return rescale(__UpperCamelCase , scale=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" return normalize(__UpperCamelCase , mean=__UpperCamelCase , std=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = ChannelDimension.FIRST , **__UpperCamelCase , ): """simple docstring""" UpperCamelCase_ = do_resize if do_resize is not None else self.do_resize UpperCamelCase_ = size if size is not None else self.size UpperCamelCase_ = get_size_dict(__UpperCamelCase , param_name="""size""" , default_to_square=__UpperCamelCase ) UpperCamelCase_ = resample if resample is not None else self.resample UpperCamelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase_ = crop_size if crop_size is not None else self.crop_size UpperCamelCase_ = get_size_dict(__UpperCamelCase , param_name="""crop_size""" , default_to_square=__UpperCamelCase ) UpperCamelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase_ = image_mean if image_mean is not None else self.image_mean UpperCamelCase_ = image_std if image_std is not None else self.image_std UpperCamelCase_ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCamelCase_ = make_list_of_images(__UpperCamelCase ) if not valid_images(__UpperCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCamelCase_ = [convert_to_rgb(__UpperCamelCase ) for image in images] # All transformations expect numpy arrays. UpperCamelCase_ = [to_numpy_array(__UpperCamelCase ) for image in images] if do_resize: UpperCamelCase_ = [self.resize(image=__UpperCamelCase , size=__UpperCamelCase , resample=__UpperCamelCase ) for image in images] if do_center_crop: UpperCamelCase_ = [self.center_crop(image=__UpperCamelCase , size=__UpperCamelCase ) for image in images] if do_rescale: UpperCamelCase_ = [self.rescale(image=__UpperCamelCase , scale=__UpperCamelCase ) for image in images] if do_normalize: UpperCamelCase_ = [self.normalize(image=__UpperCamelCase , mean=__UpperCamelCase , std=__UpperCamelCase ) for image in images] UpperCamelCase_ = [to_channel_dimension_format(__UpperCamelCase , __UpperCamelCase ) for image in images] UpperCamelCase_ = {"""pixel_values""": images} return BatchFeature(data=__UpperCamelCase , tensor_type=__UpperCamelCase )
122
1
import random class __lowerCamelCase : '''simple docstring''' @staticmethod def _UpperCAmelCase ( __UpperCAmelCase ) -> tuple[list[int], list[int]]: _a = [ord(__UpperCAmelCase ) for i in text] _a = [] _a = [] for i in plain: _a = random.randint(1 , 300 ) _a = (i + k) * k cipher.append(__UpperCAmelCase ) key.append(__UpperCAmelCase ) return cipher, key @staticmethod def _UpperCAmelCase ( __UpperCAmelCase , __UpperCAmelCase ) -> str: _a = [] for i in range(len(__UpperCAmelCase ) ): _a = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(__UpperCAmelCase ) ) return "".join(__UpperCAmelCase ) if __name__ == "__main__": __snake_case ,__snake_case = Onepad().encrypt('''Hello''') print(c, k) print(Onepad().decrypt(c, k))
353
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def A_ ( _lowerCAmelCase : str="ro", _lowerCAmelCase : Optional[Any]="en", _lowerCAmelCase : Union[str, Any]="wmt16", _lowerCAmelCase : int=None ): """simple docstring""" try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError('''run pip install datasets''' ) _a = f'{src_lang}-{tgt_lang}' print(f'Converting {dataset}-{pair}' ) _a = datasets.load_dataset(_lowerCAmelCase, _lowerCAmelCase ) if save_dir is None: _a = f'{dataset}-{pair}' _a = Path(_lowerCAmelCase ) save_dir.mkdir(exist_ok=_lowerCAmelCase ) for split in ds.keys(): print(f'Splitting {split} with {ds[split].num_rows} records' ) # to save to val.source, val.target like summary datasets _a = '''val''' if split == '''validation''' else split _a = save_dir.joinpath(f'{fn}.source' ) _a = save_dir.joinpath(f'{fn}.target' ) _a = src_path.open('''w+''' ) _a = tgt_path.open('''w+''' ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): _a = x['''translation'''] src_fp.write(ex[src_lang] + '''\n''' ) tgt_fp.write(ex[tgt_lang] + '''\n''' ) print(f'Saved {dataset} dataset to {save_dir}' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
153
0