code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness __SCREAMING_SNAKE_CASE : Optional[int] = '''\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } ''' __SCREAMING_SNAKE_CASE : Any = '''\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). ''' __SCREAMING_SNAKE_CASE : int = ''' Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric("code_eval") >>> test_cases = ["assert add(2,3)==5"] >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {\'pass@1\': 0.5, \'pass@2\': 1.0} ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' ################################################################################ !!!WARNING!!! ################################################################################ The "code_eval" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this with: >>> import os >>> os.environ["HF_ALLOW_CODE_EVAL"] = "1" ################################################################################\ ''' __SCREAMING_SNAKE_CASE : Union[str, Any] = '''The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Value('''string''' ), } ) , homepage='''https://github.com/openai/human-eval''' , codebase_urls=['''https://github.com/openai/human-eval'''] , reference_urls=['''https://github.com/openai/human-eval'''] , license=_LICENSE , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=[1, 1_0, 1_0_0] , lowerCamelCase__=4 , lowerCamelCase__=3.0 ): if os.getenv('''HF_ALLOW_CODE_EVAL''' , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError('''This metric is currently not supported on Windows.''' ) with ThreadPoolExecutor(max_workers=lowerCamelCase__ ) as executor: _lowerCamelCase = [] _lowerCamelCase = Counter() _lowerCamelCase = 0 _lowerCamelCase = defaultdict(lowerCamelCase__ ) for task_id, (candidates, test_case) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): for candidate in candidates: _lowerCamelCase = candidate + '''\n''' + test_case _lowerCamelCase = (test_program, timeout, task_id, completion_id[task_id]) _lowerCamelCase = executor.submit(lowerCamelCase__ , *lowerCamelCase__ ) futures.append(lowerCamelCase__ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(lowerCamelCase__ ): _lowerCamelCase = future.result() results[result["task_id"]].append((result['''completion_id'''], result) ) _lowerCamelCase , _lowerCamelCase = [], [] for result in results.values(): result.sort() _lowerCamelCase = [r[1]['''passed'''] for r in result] total.append(len(lowerCamelCase__ ) ) correct.append(sum(lowerCamelCase__ ) ) _lowerCamelCase = np.array(lowerCamelCase__ ) _lowerCamelCase = np.array(lowerCamelCase__ ) _lowerCamelCase = k _lowerCamelCase = {F"""pass@{k}""": estimate_pass_at_k(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : List[Any] ) -> Tuple: def estimator(lowercase_ : int , lowercase_ : int , lowercase_ : int ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(lowercase_ , lowercase_ ): _lowerCamelCase = itertools.repeat(lowercase_ , len(lowercase_ ) ) else: assert len(lowercase_ ) == len(lowercase_ ) _lowerCamelCase = iter(lowercase_ ) return np.array([estimator(int(lowercase_ ) , int(lowercase_ ) , lowercase_ ) for n, c in zip(lowercase_ , lowercase_ )] )
717
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : int = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
718
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
0
"""simple docstring""" from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : torch.FloatTensor class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__=3 , lowerCamelCase__=3 , lowerCamelCase__=("DownEncoderBlock2D",) , lowerCamelCase__=(6_4,) , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__="silu" , lowerCamelCase__=True , ): super().__init__() _lowerCamelCase = layers_per_block _lowerCamelCase = torch.nn.Convad( lowerCamelCase__ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) # down _lowerCamelCase = block_out_channels[0] for i, down_block_type in enumerate(lowerCamelCase__ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] _lowerCamelCase = i == len(lowerCamelCase__ ) - 1 _lowerCamelCase = get_down_block( lowerCamelCase__ , num_layers=self.layers_per_block , in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , add_downsample=not is_final_block , resnet_eps=1e-6 , downsample_padding=0 , resnet_act_fn=lowerCamelCase__ , resnet_groups=lowerCamelCase__ , attention_head_dim=lowerCamelCase__ , temb_channels=lowerCamelCase__ , ) self.down_blocks.append(lowerCamelCase__ ) # mid _lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=lowerCamelCase__ , output_scale_factor=1 , resnet_time_scale_shift='''default''' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCamelCase__ , temb_channels=lowerCamelCase__ , ) # out _lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCamelCase__ , eps=1e-6 ) _lowerCamelCase = nn.SiLU() _lowerCamelCase = 2 * out_channels if double_z else out_channels _lowerCamelCase = nn.Convad(block_out_channels[-1] , lowerCamelCase__ , 3 , padding=1 ) _lowerCamelCase = False def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = x _lowerCamelCase = self.conv_in(lowerCamelCase__ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCamelCase__ ): def custom_forward(*lowerCamelCase__ ): return module(*lowerCamelCase__ ) return custom_forward # down if is_torch_version('''>=''' , '''1.11.0''' ): for down_block in self.down_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) else: for down_block in self.down_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ ) # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCamelCase__ ) else: # down for down_block in self.down_blocks: _lowerCamelCase = down_block(lowerCamelCase__ ) # middle _lowerCamelCase = self.mid_block(lowerCamelCase__ ) # post-process _lowerCamelCase = self.conv_norm_out(lowerCamelCase__ ) _lowerCamelCase = self.conv_act(lowerCamelCase__ ) _lowerCamelCase = self.conv_out(lowerCamelCase__ ) return sample class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__=3 , lowerCamelCase__=3 , lowerCamelCase__=("UpDecoderBlock2D",) , lowerCamelCase__=(6_4,) , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__="silu" , lowerCamelCase__="group" , ): super().__init__() _lowerCamelCase = layers_per_block _lowerCamelCase = nn.Convad( lowerCamelCase__ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = in_channels if norm_type == '''spatial''' else None # mid _lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=lowerCamelCase__ , output_scale_factor=1 , resnet_time_scale_shift='''default''' if norm_type == '''group''' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCamelCase__ , temb_channels=lowerCamelCase__ , ) # up _lowerCamelCase = list(reversed(lowerCamelCase__ ) ) _lowerCamelCase = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCamelCase__ ): _lowerCamelCase = output_channel _lowerCamelCase = reversed_block_out_channels[i] _lowerCamelCase = i == len(lowerCamelCase__ ) - 1 _lowerCamelCase = get_up_block( lowerCamelCase__ , num_layers=self.layers_per_block + 1 , in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , prev_output_channel=lowerCamelCase__ , add_upsample=not is_final_block , resnet_eps=1e-6 , resnet_act_fn=lowerCamelCase__ , resnet_groups=lowerCamelCase__ , attention_head_dim=lowerCamelCase__ , temb_channels=lowerCamelCase__ , resnet_time_scale_shift=lowerCamelCase__ , ) self.up_blocks.append(lowerCamelCase__ ) _lowerCamelCase = output_channel # out if norm_type == "spatial": _lowerCamelCase = SpatialNorm(block_out_channels[0] , lowerCamelCase__ ) else: _lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCamelCase__ , eps=1e-6 ) _lowerCamelCase = nn.SiLU() _lowerCamelCase = nn.Convad(block_out_channels[0] , lowerCamelCase__ , 3 , padding=1 ) _lowerCamelCase = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = z _lowerCamelCase = self.conv_in(lowerCamelCase__ ) _lowerCamelCase = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCamelCase__ ): def custom_forward(*lowerCamelCase__ ): return module(*lowerCamelCase__ ) return custom_forward if is_torch_version('''>=''' , '''1.11.0''' ): # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase__ , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) _lowerCamelCase = sample.to(lowerCamelCase__ ) # up for up_block in self.up_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) else: # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = sample.to(lowerCamelCase__ ) # up for up_block in self.up_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ , lowerCamelCase__ ) else: # middle _lowerCamelCase = self.mid_block(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = sample.to(lowerCamelCase__ ) # up for up_block in self.up_blocks: _lowerCamelCase = up_block(lowerCamelCase__ , lowerCamelCase__ ) # post-process if latent_embeds is None: _lowerCamelCase = self.conv_norm_out(lowerCamelCase__ ) else: _lowerCamelCase = self.conv_norm_out(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.conv_act(lowerCamelCase__ ) _lowerCamelCase = self.conv_out(lowerCamelCase__ ) return sample class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__="random" , lowerCamelCase__=False , lowerCamelCase__=True ): super().__init__() _lowerCamelCase = n_e _lowerCamelCase = vq_embed_dim _lowerCamelCase = beta _lowerCamelCase = legacy _lowerCamelCase = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _lowerCamelCase = remap if self.remap is not None: self.register_buffer('''used''' , torch.tensor(np.load(self.remap ) ) ) _lowerCamelCase = self.used.shape[0] _lowerCamelCase = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _lowerCamelCase = self.re_embed _lowerCamelCase = self.re_embed + 1 print( F"""Remapping {self.n_e} indices to {self.re_embed} indices. """ F"""Using {self.unknown_index} for unknown indices.""" ) else: _lowerCamelCase = n_e _lowerCamelCase = sane_index_shape def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = inds.shape assert len(lowerCamelCase__ ) > 1 _lowerCamelCase = inds.reshape(ishape[0] , -1 ) _lowerCamelCase = self.used.to(lowerCamelCase__ ) _lowerCamelCase = (inds[:, :, None] == used[None, None, ...]).long() _lowerCamelCase = match.argmax(-1 ) _lowerCamelCase = match.sum(2 ) < 1 if self.unknown_index == "random": _lowerCamelCase = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _lowerCamelCase = self.unknown_index return new.reshape(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = inds.shape assert len(lowerCamelCase__ ) > 1 _lowerCamelCase = inds.reshape(ishape[0] , -1 ) _lowerCamelCase = self.used.to(lowerCamelCase__ ) if self.re_embed > self.used.shape[0]: # extra token _lowerCamelCase = 0 # simply set to zero _lowerCamelCase = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCamelCase__ ) return back.reshape(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): # reshape z -> (batch, height, width, channel) and flatten _lowerCamelCase = z.permute(0 , 2 , 3 , 1 ).contiguous() _lowerCamelCase = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _lowerCamelCase = torch.argmin(torch.cdist(lowerCamelCase__ , self.embedding.weight ) , dim=1 ) _lowerCamelCase = self.embedding(lowerCamelCase__ ).view(z.shape ) _lowerCamelCase = None _lowerCamelCase = None # compute loss for embedding if not self.legacy: _lowerCamelCase = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _lowerCamelCase = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _lowerCamelCase = z + (z_q - z).detach() # reshape back to match original input shape _lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _lowerCamelCase = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _lowerCamelCase = self.remap_to_used(lowerCamelCase__ ) _lowerCamelCase = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _lowerCamelCase = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): # shape specifying (batch, height, width, channel) if self.remap is not None: _lowerCamelCase = indices.reshape(shape[0] , -1 ) # add batch axis _lowerCamelCase = self.unmap_to_all(lowerCamelCase__ ) _lowerCamelCase = indices.reshape(-1 ) # flatten again # get quantized latent vectors _lowerCamelCase = self.embedding(lowerCamelCase__ ) if shape is not None: _lowerCamelCase = z_q.view(lowerCamelCase__ ) # reshape back to match original input shape _lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = parameters _lowerCamelCase , _lowerCamelCase = torch.chunk(lowerCamelCase__ , 2 , dim=1 ) _lowerCamelCase = torch.clamp(self.logvar , -3_0.0 , 2_0.0 ) _lowerCamelCase = deterministic _lowerCamelCase = torch.exp(0.5 * self.logvar ) _lowerCamelCase = torch.exp(self.logvar ) if self.deterministic: _lowerCamelCase = _lowerCamelCase = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def snake_case__ ( self , lowerCamelCase__ = None ): # make sure sample is on the same device as the parameters and has same dtype _lowerCamelCase = randn_tensor( self.mean.shape , generator=lowerCamelCase__ , device=self.parameters.device , dtype=self.parameters.dtype ) _lowerCamelCase = self.mean + self.std * sample return x def snake_case__ ( self , lowerCamelCase__=None ): if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=[1, 2, 3] ): if self.deterministic: return torch.Tensor([0.0] ) _lowerCamelCase = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCamelCase__ ) def snake_case__ ( self ): return self.mean
719
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : Tuple = { '''configuration_transfo_xl''': ['''TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TransfoXLConfig'''], '''tokenization_transfo_xl''': ['''TransfoXLCorpus''', '''TransfoXLTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AdaptiveEmbedding''', '''TransfoXLForSequenceClassification''', '''TransfoXLLMHeadModel''', '''TransfoXLModel''', '''TransfoXLPreTrainedModel''', '''load_tf_weights_in_transfo_xl''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFAdaptiveEmbedding''', '''TFTransfoXLForSequenceClassification''', '''TFTransfoXLLMHeadModel''', '''TFTransfoXLMainLayer''', '''TFTransfoXLModel''', '''TFTransfoXLPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
720
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
0
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
721
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Any = {'''configuration_unispeech''': ['''UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''UniSpeechConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[str] = [ '''UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST''', '''UniSpeechForCTC''', '''UniSpeechForPreTraining''', '''UniSpeechForSequenceClassification''', '''UniSpeechModel''', '''UniSpeechPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
700
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
0
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int ) -> int: return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def lowerCAmelCase_( lowercase_ : int ) -> bool: _lowerCamelCase = 0 _lowerCamelCase = number while duplicate > 0: _lowerCamelCase , _lowerCamelCase = divmod(lowercase_ , 10 ) fact_sum += factorial(lowercase_ ) return fact_sum == number if __name__ == "__main__": print('''Program to check whether a number is a Krisnamurthy Number or not.''') __SCREAMING_SNAKE_CASE : Optional[int] = int(input('''Enter number: ''').strip()) print( F"""{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.""" )
701
"""simple docstring""" from typing import Any class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) _lowerCamelCase = temp.next print() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = Node(lowerCamelCase__ ) _lowerCamelCase = self.head _lowerCamelCase = new_node def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if node_data_a == node_data_a: return else: _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next if node_a is None or node_a is None: return _lowerCamelCase , _lowerCamelCase = node_a.data, node_a.data if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
623
0
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np __SCREAMING_SNAKE_CASE : List[Any] = re.compile(R'''\b(a|an|the)\b''', re.UNICODE) __SCREAMING_SNAKE_CASE : Optional[Any] = None def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = argparse.ArgumentParser('''Official evaluation script for SQuAD version 2.0.''' ) parser.add_argument('''data_file''' , metavar='''data.json''' , help='''Input data JSON file.''' ) parser.add_argument('''pred_file''' , metavar='''pred.json''' , help='''Model predictions.''' ) parser.add_argument( '''--out-file''' , '''-o''' , metavar='''eval.json''' , help='''Write accuracy metrics to file (default is stdout).''' ) parser.add_argument( '''--na-prob-file''' , '''-n''' , metavar='''na_prob.json''' , help='''Model estimates of probability of no answer.''' ) parser.add_argument( '''--na-prob-thresh''' , '''-t''' , type=lowercase_ , default=1.0 , help='''Predict "" if no-answer probability exceeds this (default = 1.0).''' , ) parser.add_argument( '''--out-image-dir''' , '''-p''' , metavar='''out_images''' , default=lowercase_ , help='''Save precision-recall curves to directory.''' ) parser.add_argument('''--verbose''' , '''-v''' , action='''store_true''' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def lowerCAmelCase_( lowercase_ : int ) -> List[str]: _lowerCamelCase = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: _lowerCamelCase = bool(qa['''answers''']['''text'''] ) return qid_to_has_ans def lowerCAmelCase_( lowercase_ : str ) -> List[str]: def remove_articles(lowercase_ : List[Any] ): return ARTICLES_REGEX.sub(''' ''' , lowercase_ ) def white_space_fix(lowercase_ : int ): return " ".join(text.split() ) def remove_punc(lowercase_ : List[str] ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Tuple ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : int ) -> List[str]: if not s: return [] return normalize_answer(lowercase_ ).split() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : Optional[Any] ) -> Any: return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Any ) -> List[str]: _lowerCamelCase = get_tokens(lowercase_ ) _lowerCamelCase = get_tokens(lowercase_ ) _lowerCamelCase = collections.Counter(lowercase_ ) & collections.Counter(lowercase_ ) _lowerCamelCase = sum(common.values() ) if len(lowercase_ ) == 0 or len(lowercase_ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 _lowerCamelCase = 1.0 * num_same / len(lowercase_ ) _lowerCamelCase = 1.0 * num_same / len(lowercase_ ) _lowerCamelCase = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase_( lowercase_ : int , lowercase_ : Any ) -> List[str]: _lowerCamelCase = {} _lowerCamelCase = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: _lowerCamelCase = qa['''id'''] _lowerCamelCase = [t for t in qa['''answers''']['''text'''] if normalize_answer(lowercase_ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string _lowerCamelCase = [''''''] if qid not in preds: print(F"""Missing prediction for {qid}""" ) continue _lowerCamelCase = preds[qid] # Take max over all gold answers _lowerCamelCase = max(compute_exact(lowercase_ , lowercase_ ) for a in gold_answers ) _lowerCamelCase = max(compute_fa(lowercase_ , lowercase_ ) for a in gold_answers ) return exact_scores, fa_scores def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : List[str] , lowercase_ : List[Any] , lowercase_ : Optional[int] ) -> Union[str, Any]: _lowerCamelCase = {} for qid, s in scores.items(): _lowerCamelCase = na_probs[qid] > na_prob_thresh if pred_na: _lowerCamelCase = float(not qid_to_has_ans[qid] ) else: _lowerCamelCase = s return new_scores def lowerCAmelCase_( lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[Any]=None ) -> Any: if not qid_list: _lowerCamelCase = len(lowercase_ ) return collections.OrderedDict( [ ('''exact''', 1_00.0 * sum(exact_scores.values() ) / total), ('''f1''', 1_00.0 * sum(fa_scores.values() ) / total), ('''total''', total), ] ) else: _lowerCamelCase = len(lowercase_ ) return collections.OrderedDict( [ ('''exact''', 1_00.0 * sum(exact_scores[k] for k in qid_list ) / total), ('''f1''', 1_00.0 * sum(fa_scores[k] for k in qid_list ) / total), ('''total''', total), ] ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Tuple , lowercase_ : Any ) -> List[Any]: for k in new_eval: _lowerCamelCase = new_eval[k] def lowerCAmelCase_( lowercase_ : int , lowercase_ : Tuple , lowercase_ : Tuple , lowercase_ : List[Any] ) -> Optional[Any]: plt.step(lowercase_ , lowercase_ , color='''b''' , alpha=0.2 , where='''post''' ) plt.fill_between(lowercase_ , lowercase_ , step='''post''' , alpha=0.2 , color='''b''' ) plt.xlabel('''Recall''' ) plt.ylabel('''Precision''' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(lowercase_ ) plt.savefig(lowercase_ ) plt.clf() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int , lowercase_ : Union[str, Any] , lowercase_ : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Any=None ) -> Union[str, Any]: _lowerCamelCase = sorted(lowercase_ , key=lambda lowercase_ : na_probs[k] ) _lowerCamelCase = 0.0 _lowerCamelCase = 1.0 _lowerCamelCase = 0.0 _lowerCamelCase = [1.0] _lowerCamelCase = [0.0] _lowerCamelCase = 0.0 for i, qid in enumerate(lowercase_ ): if qid_to_has_ans[qid]: true_pos += scores[qid] _lowerCamelCase = true_pos / float(i + 1 ) _lowerCamelCase = true_pos / float(lowercase_ ) if i == len(lowercase_ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(lowercase_ ) recalls.append(lowercase_ ) if out_image: plot_pr_curve(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) return {"ap": 1_00.0 * avg_prec} def lowerCAmelCase_( lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : Any , lowercase_ : Optional[int] , lowercase_ : Dict , lowercase_ : Union[str, Any] ) -> str: if out_image_dir and not os.path.exists(lowercase_ ): os.makedirs(lowercase_ ) _lowerCamelCase = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return _lowerCamelCase = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , '''pr_exact.png''' ) , title='''Precision-Recall curve for Exact Match score''' , ) _lowerCamelCase = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , '''pr_f1.png''' ) , title='''Precision-Recall curve for F1 score''' , ) _lowerCamelCase = {k: float(lowercase_ ) for k, v in qid_to_has_ans.items()} _lowerCamelCase = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , '''pr_oracle.png''' ) , title='''Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)''' , ) merge_eval(lowercase_ , lowercase_ , '''pr_exact''' ) merge_eval(lowercase_ , lowercase_ , '''pr_f1''' ) merge_eval(lowercase_ , lowercase_ , '''pr_oracle''' ) def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : Dict , lowercase_ : Union[str, Any] , lowercase_ : List[Any] ) -> Tuple: if not qid_list: return _lowerCamelCase = [na_probs[k] for k in qid_list] _lowerCamelCase = np.ones_like(lowercase_ ) / float(len(lowercase_ ) ) plt.hist(lowercase_ , weights=lowercase_ , bins=20 , range=(0.0, 1.0) ) plt.xlabel('''Model probability of no-answer''' ) plt.ylabel('''Proportion of dataset''' ) plt.title(F"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(lowercase_ , F"""na_prob_hist_{name}.png""" ) ) plt.clf() def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Dict , lowercase_ : Dict , lowercase_ : List[Any] ) -> Optional[Any]: _lowerCamelCase = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) _lowerCamelCase = num_no_ans _lowerCamelCase = cur_score _lowerCamelCase = 0.0 _lowerCamelCase = sorted(lowercase_ , key=lambda lowercase_ : na_probs[k] ) for i, qid in enumerate(lowercase_ ): if qid not in scores: continue if qid_to_has_ans[qid]: _lowerCamelCase = scores[qid] else: if preds[qid]: _lowerCamelCase = -1 else: _lowerCamelCase = 0 cur_score += diff if cur_score > best_score: _lowerCamelCase = cur_score _lowerCamelCase = na_probs[qid] return 1_00.0 * best_score / len(lowercase_ ), best_thresh def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Tuple , lowercase_ : Optional[Any] , lowercase_ : Dict , lowercase_ : Tuple , lowercase_ : Optional[Any] ) -> Dict: _lowerCamelCase , _lowerCamelCase = find_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase , _lowerCamelCase = find_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = best_exact _lowerCamelCase = exact_thresh _lowerCamelCase = best_fa _lowerCamelCase = fa_thresh def lowerCAmelCase_( ) -> Tuple: with open(OPTS.data_file ) as f: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = dataset_json['''data'''] with open(OPTS.pred_file ) as f: _lowerCamelCase = json.load(lowercase_ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: _lowerCamelCase = json.load(lowercase_ ) else: _lowerCamelCase = {k: 0.0 for k in preds} _lowerCamelCase = make_qid_to_has_ans(lowercase_ ) # maps qid to True/False _lowerCamelCase = [k for k, v in qid_to_has_ans.items() if v] _lowerCamelCase = [k for k, v in qid_to_has_ans.items() if not v] _lowerCamelCase , _lowerCamelCase = get_raw_scores(lowercase_ , lowercase_ ) _lowerCamelCase = apply_no_ans_threshold(lowercase_ , lowercase_ , lowercase_ , OPTS.na_prob_thresh ) _lowerCamelCase = apply_no_ans_threshold(lowercase_ , lowercase_ , lowercase_ , OPTS.na_prob_thresh ) _lowerCamelCase = make_eval_dict(lowercase_ , lowercase_ ) if has_ans_qids: _lowerCamelCase = make_eval_dict(lowercase_ , lowercase_ , qid_list=lowercase_ ) merge_eval(lowercase_ , lowercase_ , '''HasAns''' ) if no_ans_qids: _lowerCamelCase = make_eval_dict(lowercase_ , lowercase_ , qid_list=lowercase_ ) merge_eval(lowercase_ , lowercase_ , '''NoAns''' ) if OPTS.na_prob_file: find_all_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , OPTS.out_image_dir ) histogram_na_prob(lowercase_ , lowercase_ , OPTS.out_image_dir , '''hasAns''' ) histogram_na_prob(lowercase_ , lowercase_ , OPTS.out_image_dir , '''noAns''' ) if OPTS.out_file: with open(OPTS.out_file , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) else: print(json.dumps(lowercase_ , indent=2 ) ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('''Agg''') import matplotlib.pyplot as plt main()
702
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __SCREAMING_SNAKE_CASE : Optional[Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[Any]: from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> List[str]: from diffusers.utils.testing_utils import pytest_terminal_summary_main _lowerCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
623
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[str] = { '''microsoft/beit-base-patch16-224-pt22k''': ( '''https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json''' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'beit' def __init__( self , lowerCamelCase__=8_1_9_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=2_2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=3 , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=True , lowerCamelCase__=[3, 5, 7, 1_1] , lowerCamelCase__=[1, 2, 3, 6] , lowerCamelCase__=True , lowerCamelCase__=0.4 , lowerCamelCase__=2_5_6 , lowerCamelCase__=1 , lowerCamelCase__=False , lowerCamelCase__=2_5_5 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = use_mask_token _lowerCamelCase = use_absolute_position_embeddings _lowerCamelCase = use_relative_position_bias _lowerCamelCase = use_shared_relative_position_bias _lowerCamelCase = layer_scale_init_value _lowerCamelCase = drop_path_rate _lowerCamelCase = use_mean_pooling # decode head attributes (semantic segmentation) _lowerCamelCase = out_indices _lowerCamelCase = pool_scales # auxiliary head attributes (semantic segmentation) _lowerCamelCase = use_auxiliary_head _lowerCamelCase = auxiliary_loss_weight _lowerCamelCase = auxiliary_channels _lowerCamelCase = auxiliary_num_convs _lowerCamelCase = auxiliary_concat_input _lowerCamelCase = semantic_loss_ignore_index class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Dict = version.parse('1.11' ) @property def snake_case__ ( self ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def snake_case__ ( self ): return 1e-4
703
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
0
"""simple docstring""" import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset __SCREAMING_SNAKE_CASE = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__() _lowerCamelCase = torchvision.models.resnetaaa(pretrained=lowerCamelCase__ ) _lowerCamelCase = list(model.children() )[:-2] _lowerCamelCase = nn.Sequential(*lowerCamelCase__ ) _lowerCamelCase = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def snake_case__ ( self , lowerCamelCase__ ): # Bx3x224x224 -> Bx2048x7x7 -> Bx2048xN -> BxNx2048 _lowerCamelCase = self.pool(self.model(lowerCamelCase__ ) ) _lowerCamelCase = torch.flatten(lowerCamelCase__ , start_dim=2 ) _lowerCamelCase = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = [json.loads(lowerCamelCase__ ) for l in open(lowerCamelCase__ )] _lowerCamelCase = os.path.dirname(lowerCamelCase__ ) _lowerCamelCase = tokenizer _lowerCamelCase = labels _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = max_seq_length _lowerCamelCase = transforms def __len__( self ): return len(self.data ) def __getitem__( self , lowerCamelCase__ ): _lowerCamelCase = torch.LongTensor(self.tokenizer.encode(self.data[index]['''text'''] , add_special_tokens=lowerCamelCase__ ) ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = sentence[0], sentence[1:-1], sentence[-1] _lowerCamelCase = sentence[: self.max_seq_length] _lowerCamelCase = torch.zeros(self.n_classes ) _lowerCamelCase = 1 _lowerCamelCase = Image.open(os.path.join(self.data_dir , self.data[index]['''img'''] ) ).convert('''RGB''' ) _lowerCamelCase = self.transforms(lowerCamelCase__ ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def snake_case__ ( self ): _lowerCamelCase = Counter() for row in self.data: label_freqs.update(row['''label'''] ) return label_freqs def lowerCAmelCase_( lowercase_ : Any ) -> Any: _lowerCamelCase = [len(row['''sentence'''] ) for row in batch] _lowerCamelCase , _lowerCamelCase = len(lowercase_ ), max(lowercase_ ) _lowerCamelCase = torch.zeros(lowercase_ , lowercase_ , dtype=torch.long ) _lowerCamelCase = torch.zeros(lowercase_ , lowercase_ , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(lowercase_ , lowercase_ ) ): _lowerCamelCase = input_row['''sentence'''] _lowerCamelCase = 1 _lowerCamelCase = torch.stack([row['''image'''] for row in batch] ) _lowerCamelCase = torch.stack([row['''label'''] for row in batch] ) _lowerCamelCase = torch.stack([row['''image_start_token'''] for row in batch] ) _lowerCamelCase = torch.stack([row['''image_end_token'''] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def lowerCAmelCase_( ) -> Union[str, Any]: return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def lowerCAmelCase_( ) -> List[str]: return transforms.Compose( [ transforms.Resize(2_56 ), transforms.CenterCrop(2_24 ), transforms.ToTensor(), transforms.Normalize( mean=[0.4_6_7_7_7_0_4_4, 0.4_4_5_3_1_4_2_9, 0.4_0_6_6_1_0_1_7] , std=[0.1_2_2_2_1_9_9_4, 0.1_2_1_4_5_8_3_5, 0.1_4_3_8_0_4_6_9] , ), ] )
704
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
0
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : str , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : str=True , lowercase_ : Tuple="pt" ) -> Union[str, Any]: _lowerCamelCase = {'''add_prefix_space''': True} if isinstance(lowercase_ , lowercase_ ) and not line.startswith(''' ''' ) else {} _lowerCamelCase = padding_side return tokenizer( [line] , max_length=lowercase_ , padding='''max_length''' if pad_to_max_length else None , truncation=lowercase_ , return_tensors=lowercase_ , add_special_tokens=lowercase_ , **lowercase_ , ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Any , lowercase_ : List[Any]=None , ) -> Union[str, Any]: _lowerCamelCase = input_ids.ne(lowercase_ ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="train" , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__="" , ): super().__init__() _lowerCamelCase = Path(lowerCamelCase__ ).joinpath(type_path + '''.source''' ) _lowerCamelCase = Path(lowerCamelCase__ ).joinpath(type_path + '''.target''' ) _lowerCamelCase = self.get_char_lens(self.src_file ) _lowerCamelCase = max_source_length _lowerCamelCase = max_target_length assert min(self.src_lens ) > 0, F"""found empty line in {self.src_file}""" _lowerCamelCase = tokenizer _lowerCamelCase = prefix if n_obs is not None: _lowerCamelCase = self.src_lens[:n_obs] _lowerCamelCase = src_lang _lowerCamelCase = tgt_lang def __len__( self ): return len(self.src_lens ) def __getitem__( self , lowerCamelCase__ ): _lowerCamelCase = index + 1 # linecache starts at 1 _lowerCamelCase = self.prefix + linecache.getline(str(self.src_file ) , lowerCamelCase__ ).rstrip('''\n''' ) _lowerCamelCase = linecache.getline(str(self.tgt_file ) , lowerCamelCase__ ).rstrip('''\n''' ) assert source_line, F"""empty source line for index {index}""" assert tgt_line, F"""empty tgt line for index {index}""" # Need to add eos token manually for T5 if isinstance(self.tokenizer , lowerCamelCase__ ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right _lowerCamelCase = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , lowerCamelCase__ ) else self.tokenizer ) _lowerCamelCase = self.tokenizer.generator if isinstance(self.tokenizer , lowerCamelCase__ ) else self.tokenizer _lowerCamelCase = encode_line(lowerCamelCase__ , lowerCamelCase__ , self.max_source_length , '''right''' ) _lowerCamelCase = encode_line(lowerCamelCase__ , lowerCamelCase__ , self.max_target_length , '''right''' ) _lowerCamelCase = source_inputs['''input_ids'''].squeeze() _lowerCamelCase = target_inputs['''input_ids'''].squeeze() _lowerCamelCase = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def snake_case__ ( lowerCamelCase__ ): return [len(lowerCamelCase__ ) for x in Path(lowerCamelCase__ ).open().readlines()] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = torch.stack([x['''input_ids'''] for x in batch] ) _lowerCamelCase = torch.stack([x['''attention_mask'''] for x in batch] ) _lowerCamelCase = torch.stack([x['''decoder_input_ids'''] for x in batch] ) _lowerCamelCase = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , lowerCamelCase__ ) else self.tokenizer.pad_token_id ) _lowerCamelCase = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , lowerCamelCase__ ) else self.tokenizer.pad_token_id ) _lowerCamelCase = trim_batch(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = trim_batch(lowerCamelCase__ , lowerCamelCase__ , attention_mask=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch __SCREAMING_SNAKE_CASE : Union[str, Any] = getLogger(__name__) def lowerCAmelCase_( lowercase_ : List[List] ) -> Any: return list(itertools.chain.from_iterable(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> None: _lowerCamelCase = get_git_info() save_json(lowercase_ , os.path.join(lowercase_ , '''git_log.json''' ) ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : Union[str, Any] , lowercase_ : Any=4 , **lowercase_ : Union[str, Any] ) -> Union[str, Any]: with open(lowercase_ , '''w''' ) as f: json.dump(lowercase_ , lowercase_ , indent=lowercase_ , **lowercase_ ) def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> List[str]: with open(lowercase_ ) as f: return json.load(lowercase_ ) def lowerCAmelCase_( ) -> Tuple: _lowerCamelCase = git.Repo(search_parent_directories=lowercase_ ) _lowerCamelCase = { '''repo_id''': str(lowercase_ ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def lowerCAmelCase_( lowercase_ : Callable , lowercase_ : Iterable ) -> List: return list(map(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : List[str] ) -> int: with open(lowercase_ , '''wb''' ) as f: return pickle.dump(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : List[Any] ) -> List[Any]: def remove_articles(lowercase_ : str ): return re.sub(r'''\b(a|an|the)\b''' , ''' ''' , lowercase_ ) def white_space_fix(lowercase_ : Dict ): return " ".join(text.split() ) def remove_punc(lowercase_ : Dict ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : List[str] ) -> List[str]: _lowerCamelCase = normalize_answer(lowercase_ ).split() _lowerCamelCase = normalize_answer(lowercase_ ).split() _lowerCamelCase = Counter(lowercase_ ) & Counter(lowercase_ ) _lowerCamelCase = sum(common.values() ) if num_same == 0: return 0 _lowerCamelCase = 1.0 * num_same / len(lowercase_ ) _lowerCamelCase = 1.0 * num_same / len(lowercase_ ) _lowerCamelCase = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : str ) -> List[Any]: return normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : List[str] ) -> Dict: assert len(lowercase_ ) == len(lowercase_ ) _lowerCamelCase = 0 for hypo, pred in zip(lowercase_ , lowercase_ ): em += exact_match_score(lowercase_ , lowercase_ ) if len(lowercase_ ) > 0: em /= len(lowercase_ ) return {"em": em} def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> Optional[int]: return model_prefix.startswith('''rag''' ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Any , lowercase_ : str ) -> Tuple: _lowerCamelCase = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead _lowerCamelCase = '''dropout_rate''' for p in extra_params: if getattr(lowercase_ , lowercase_ , lowercase_ ): if not hasattr(lowercase_ , lowercase_ ) and not hasattr(lowercase_ , equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(lowercase_ ) ) delattr(lowercase_ , lowercase_ ) continue _lowerCamelCase = p if hasattr(lowercase_ , lowercase_ ) else equivalent_param[p] setattr(lowercase_ , lowercase_ , getattr(lowercase_ , lowercase_ ) ) delattr(lowercase_ , lowercase_ ) return hparams, config
705
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
706
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
0
"""simple docstring""" import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class lowerCamelCase_( tf.keras.optimizers.schedules.LearningRateSchedule ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): super().__init__() _lowerCamelCase = initial_learning_rate _lowerCamelCase = warmup_steps _lowerCamelCase = power _lowerCamelCase = decay_schedule_fn _lowerCamelCase = name def __call__( self , lowerCamelCase__ ): with tf.name_scope(self.name or '''WarmUp''' ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. _lowerCamelCase = tf.cast(lowerCamelCase__ , tf.floataa ) _lowerCamelCase = tf.cast(self.warmup_steps , tf.floataa ) _lowerCamelCase = global_step_float / warmup_steps_float _lowerCamelCase = self.initial_learning_rate * tf.math.pow(lowerCamelCase__ , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=lowerCamelCase__ , ) def snake_case__ ( self ): return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def lowerCAmelCase_( lowercase_ : float , lowercase_ : int , lowercase_ : int , lowercase_ : float = 0.0 , lowercase_ : float = 0.9 , lowercase_ : float = 0.9_9_9 , lowercase_ : float = 1e-8 , lowercase_ : Optional[float] = None , lowercase_ : Optional[float] = None , lowercase_ : float = 0.0 , lowercase_ : float = 1.0 , lowercase_ : Optional[List[str]] = None , ) -> List[Any]: '''simple docstring''' _lowerCamelCase = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=lowercase_ , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=lowercase_ , ) if num_warmup_steps: _lowerCamelCase = WarmUp( initial_learning_rate=lowercase_ , decay_schedule_fn=lowercase_ , warmup_steps=lowercase_ , ) if weight_decay_rate > 0.0: _lowerCamelCase = AdamWeightDecay( learning_rate=lowercase_ , weight_decay_rate=lowercase_ , beta_a=lowercase_ , beta_a=lowercase_ , epsilon=lowercase_ , clipnorm=lowercase_ , global_clipnorm=lowercase_ , exclude_from_weight_decay=['''LayerNorm''', '''layer_norm''', '''bias'''] , include_in_weight_decay=lowercase_ , ) else: _lowerCamelCase = tf.keras.optimizers.Adam( learning_rate=lowercase_ , beta_a=lowercase_ , beta_a=lowercase_ , epsilon=lowercase_ , clipnorm=lowercase_ , global_clipnorm=lowercase_ , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ = 0.0_0_1 , lowerCamelCase__ = 0.9 , lowerCamelCase__ = 0.9_9_9 , lowerCamelCase__ = 1e-7 , lowerCamelCase__ = False , lowerCamelCase__ = 0.0 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = "AdamWeightDecay" , **lowerCamelCase__ , ): super().__init__(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = weight_decay_rate _lowerCamelCase = include_in_weight_decay _lowerCamelCase = exclude_from_weight_decay @classmethod def snake_case__ ( cls , lowerCamelCase__ ): _lowerCamelCase = {'''WarmUp''': WarmUp} return super(lowerCamelCase__ , cls ).from_config(lowerCamelCase__ , custom_objects=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): super(lowerCamelCase__ , self )._prepare_local(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tf.constant( self.weight_decay_rate , name='''adam_weight_decay_rate''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]['''weight_decay_rate'''] , use_locking=self._use_locking , ) return tf.no_op() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): _lowerCamelCase , _lowerCamelCase = list(zip(*lowerCamelCase__ ) ) return super(lowerCamelCase__ , self ).apply_gradients(zip(lowerCamelCase__ , lowerCamelCase__ ) , name=lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if apply_state is None: return self._decayed_lr_t[var_dtype], {} _lowerCamelCase = apply_state or {} _lowerCamelCase = apply_state.get((var_device, var_dtype) ) if coefficients is None: _lowerCamelCase = self._fallback_apply_state(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase , _lowerCamelCase = self._get_lr(var.device , var.dtype.base_dtype , lowerCamelCase__ ) _lowerCamelCase = self._decay_weights_op(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) with tf.control_dependencies([decay] ): return super(lowerCamelCase__ , self )._resource_apply_dense(lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase , _lowerCamelCase = self._get_lr(var.device , var.dtype.base_dtype , lowerCamelCase__ ) _lowerCamelCase = self._decay_weights_op(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) with tf.control_dependencies([decay] ): return super(lowerCamelCase__ , self )._resource_apply_sparse(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = super().get_config() config.update({'''weight_decay_rate''': self.weight_decay_rate} ) return config def snake_case__ ( self , lowerCamelCase__ ): if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(lowerCamelCase__ , lowerCamelCase__ ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(lowerCamelCase__ , lowerCamelCase__ ) is not None: return False return True class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self ): _lowerCamelCase = [] _lowerCamelCase = None @property def snake_case__ ( self ): if self._accum_steps is None: _lowerCamelCase = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=lowerCamelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def snake_case__ ( self ): if not self._gradients: raise ValueError('''The accumulator should be called first to initialize the gradients''' ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self , lowerCamelCase__ ): if not self._gradients: _lowerCamelCase = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(lowerCamelCase__ ) , trainable=lowerCamelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(lowerCamelCase__ ) != len(self._gradients ): raise ValueError(F"""Expected {len(self._gradients )} gradients, but got {len(lowerCamelCase__ )}""" ) for accum_gradient, gradient in zip(self._gradients , lowerCamelCase__ ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(lowerCamelCase__ ) self._accum_steps.assign_add(1 ) def snake_case__ ( self ): if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(lowerCamelCase__ ) )
707
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
0
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device __SCREAMING_SNAKE_CASE : Any = False class lowerCamelCase_( unittest.TestCase ): pass @nightly @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): def snake_case__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self ): _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger ''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = generator.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def snake_case__ ( self ): _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger ''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=5_0 , output_type='''numpy''' ).images _lowerCamelCase = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
708
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
"""simple docstring""" import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency __SCREAMING_SNAKE_CASE : List[Any] = { '''E''': 12.70, '''T''': 9.06, '''A''': 8.17, '''O''': 7.51, '''I''': 6.97, '''N''': 6.75, '''S''': 6.33, '''H''': 6.09, '''R''': 5.99, '''D''': 4.25, '''L''': 4.03, '''C''': 2.78, '''U''': 2.76, '''M''': 2.41, '''W''': 2.36, '''F''': 2.23, '''G''': 2.02, '''Y''': 1.97, '''P''': 1.93, '''B''': 1.29, '''V''': 0.98, '''K''': 0.77, '''J''': 0.15, '''X''': 0.15, '''Q''': 0.10, '''Z''': 0.07, } __SCREAMING_SNAKE_CASE : str = '''ETAOINSHRDLCUMWFGYPBVKJXQZ''' __SCREAMING_SNAKE_CASE : Dict = '''ABCDEFGHIJKLMNOPQRSTUVWXYZ''' def lowerCAmelCase_( lowercase_ : str ) -> dict[str, int]: _lowerCamelCase = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def lowerCAmelCase_( lowercase_ : tuple ) -> str: return x[0] def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase = get_letter_count(lowercase_ ) _lowerCamelCase = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(lowercase_ ) _lowerCamelCase = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=lowercase_ ) _lowerCamelCase = ''''''.join(freq_to_letter[freq] ) _lowerCamelCase = list(freq_to_letter_str.items() ) freq_pairs.sort(key=lowercase_ , reverse=lowercase_ ) _lowerCamelCase = [freq_pair[1] for freq_pair in freq_pairs] return "".join(lowercase_ ) def lowerCAmelCase_( lowercase_ : str ) -> int: _lowerCamelCase = get_frequency_order(lowercase_ ) _lowerCamelCase = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
709
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
0
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = ['image_processor', 'tokenizer'] lowercase__ : Tuple = 'AutoImageProcessor' lowercase__ : Optional[int] = 'AutoTokenizer' def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ ): _lowerCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , lowerCamelCase__ , ) _lowerCamelCase = kwargs.pop('''feature_extractor''' ) _lowerCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.image_processor _lowerCamelCase = False def __call__( self , *lowerCamelCase__ , **lowerCamelCase__ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''images''' , lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''text''' , lowerCamelCase__ ) if len(lowerCamelCase__ ) > 0: _lowerCamelCase = args[0] _lowerCamelCase = args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: _lowerCamelCase = self.image_processor(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) if text is not None: _lowerCamelCase = self.tokenizer(lowerCamelCase__ , **lowerCamelCase__ ) if text is None: return inputs elif images is None: return encodings else: _lowerCamelCase = encodings['''input_ids'''] return inputs def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) @contextmanager def snake_case__ ( self ): warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) _lowerCamelCase = True _lowerCamelCase = self.tokenizer yield _lowerCamelCase = self.image_processor _lowerCamelCase = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=None ): if added_vocab is None: _lowerCamelCase = self.tokenizer.get_added_vocab() _lowerCamelCase = {} while tokens: _lowerCamelCase = re.search(R'''<s_(.*?)>''' , lowerCamelCase__ , re.IGNORECASE ) if start_token is None: break _lowerCamelCase = start_token.group(1 ) _lowerCamelCase = re.search(RF"""</s_{key}>""" , lowerCamelCase__ , re.IGNORECASE ) _lowerCamelCase = start_token.group() if end_token is None: _lowerCamelCase = tokens.replace(lowerCamelCase__ , '''''' ) else: _lowerCamelCase = end_token.group() _lowerCamelCase = re.escape(lowerCamelCase__ ) _lowerCamelCase = re.escape(lowerCamelCase__ ) _lowerCamelCase = re.search(F"""{start_token_escaped}(.*?){end_token_escaped}""" , lowerCamelCase__ , re.IGNORECASE ) if content is not None: _lowerCamelCase = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _lowerCamelCase = self.tokenajson(lowerCamelCase__ , is_inner_value=lowerCamelCase__ , added_vocab=lowerCamelCase__ ) if value: if len(lowerCamelCase__ ) == 1: _lowerCamelCase = value[0] _lowerCamelCase = value else: # leaf nodes _lowerCamelCase = [] for leaf in content.split(R'''<sep/>''' ): _lowerCamelCase = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _lowerCamelCase = leaf[1:-2] # for categorical special tokens output[key].append(lowerCamelCase__ ) if len(output[key] ) == 1: _lowerCamelCase = output[key][0] _lowerCamelCase = tokens[tokens.find(lowerCamelCase__ ) + len(lowerCamelCase__ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=lowerCamelCase__ , added_vocab=lowerCamelCase__ ) if len(lowerCamelCase__ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def snake_case__ ( self ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , lowerCamelCase__ , ) return self.image_processor_class @property def snake_case__ ( self ): warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , lowerCamelCase__ , ) return self.image_processor
710
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = StableDiffusionInpaintPipeline lowercase__ : Dict = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowercase__ : List[str] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowercase__ : str = frozenset([] ) def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , ) _lowerCamelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=5_1_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCamelCase = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('''RGB''' ).resize((6_4, 6_4) ) _lowerCamelCase = Image.fromarray(np.uinta(image + 4 ) ).convert('''RGB''' ).resize((6_4, 6_4) ) if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionInpaintPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _lowerCamelCase = np.array([0.4_7_2_7, 0.5_7_3_5, 0.3_9_4_1, 0.5_4_4_6, 0.5_9_2_6, 0.4_3_9_4, 0.5_0_6_2, 0.4_6_5_4, 0.4_4_7_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self ): _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) _lowerCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''' ) _lowerCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' _lowerCamelCase = StableDiffusionInpaintPipeline.from_pretrained(lowerCamelCase__ , safety_checker=lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() _lowerCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , image=lowerCamelCase__ , mask_image=lowerCamelCase__ , generator=lowerCamelCase__ , output_type='''np''' , ) _lowerCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9e-3 def snake_case__ ( self ): _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) _lowerCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''' ) _lowerCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' _lowerCamelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase__ , torch_dtype=torch.floataa , safety_checker=lowerCamelCase__ , ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() _lowerCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , image=lowerCamelCase__ , mask_image=lowerCamelCase__ , generator=lowerCamelCase__ , output_type='''np''' , ) _lowerCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5e-1 def snake_case__ ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) _lowerCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' _lowerCamelCase = PNDMScheduler.from_pretrained(lowerCamelCase__ , subfolder='''scheduler''' ) _lowerCamelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase__ , safety_checker=lowerCamelCase__ , scheduler=lowerCamelCase__ , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _lowerCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , image=lowerCamelCase__ , mask_image=lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=2 , output_type='''np''' , ) _lowerCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.6_5 * 1_0**9
711
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
0
"""simple docstring""" import argparse from collections import defaultdict import yaml __SCREAMING_SNAKE_CASE : Any = '''docs/source/en/_toctree.yml''' def lowerCAmelCase_( lowercase_ : int ) -> Tuple: _lowerCamelCase = defaultdict(lowercase_ ) for doc in model_doc: counts[doc["local"]] += 1 _lowerCamelCase = [key for key, value in counts.items() if value > 1] _lowerCamelCase = [] for duplicate_key in duplicates: _lowerCamelCase = list({doc['''title'''] for doc in model_doc if doc['''local'''] == duplicate_key} ) if len(lowercase_ ) > 1: raise ValueError( F"""{duplicate_key} is present several times in the documentation table of content at """ '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['''local''']] == 1] ) # Sort return sorted(lowercase_ , key=lambda lowercase_ : s["title"].lower() ) def lowerCAmelCase_( lowercase_ : List[Any]=False ) -> List[str]: with open(lowercase_ , encoding='''utf-8''' ) as f: _lowerCamelCase = yaml.safe_load(f.read() ) # Get to the API doc _lowerCamelCase = 0 while content[api_idx]["title"] != "API": api_idx += 1 _lowerCamelCase = content[api_idx]['''sections'''] # Then to the model doc _lowerCamelCase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 _lowerCamelCase = api_doc[model_idx]['''sections'''] _lowerCamelCase = [(idx, section) for idx, section in enumerate(lowercase_ ) if '''sections''' in section] _lowerCamelCase = False for idx, modality_doc in modalities_docs: _lowerCamelCase = modality_doc['''sections'''] _lowerCamelCase = clean_model_doc_toc(lowercase_ ) if old_modality_doc != new_modality_doc: _lowerCamelCase = True if overwrite: _lowerCamelCase = new_modality_doc if diff: if overwrite: _lowerCamelCase = model_doc _lowerCamelCase = api_doc with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowercase_ , allow_unicode=lowercase_ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') __SCREAMING_SNAKE_CASE : Dict = parser.parse_args() check_model_doc(args.fix_and_overwrite)
712
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope _lowerCamelCase = self.vocab_size - 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) _lowerCamelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = OpenAIGPTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , head_mask=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = OpenAIGPTLMHeadModel(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = OpenAIGPTDoubleHeadsModel(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = OpenAIGPTForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) lowercase__ : List[Any] = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly lowercase__ : Union[str, Any] = ( { 'feature-extraction': OpenAIGPTModel, 'text-classification': OpenAIGPTForSequenceClassification, 'text-generation': OpenAIGPTLMHeadModel, 'zero-shot': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = super()._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": _lowerCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=lowerCamelCase__ , ) _lowerCamelCase = inputs_dict['''labels'''] _lowerCamelCase = inputs_dict['''labels'''] _lowerCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=lowerCamelCase__ , ) _lowerCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase__ ) return inputs_dict def snake_case__ ( self ): _lowerCamelCase = OpenAIGPTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , n_embd=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = OpenAIGPTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = OpenAIGPTLMHeadModel.from_pretrained('''openai-gpt''' ) model.to(lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]] , dtype=torch.long , device=lowerCamelCase__ ) # the president is _lowerCamelCase = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the _lowerCamelCase = model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ ) self.assertListEqual(output_ids[0].tolist() , lowerCamelCase__ )
713
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
0
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
714
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
0
"""simple docstring""" import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = GPTaTokenizer lowercase__ : Optional[int] = GPTaTokenizerFast lowercase__ : List[str] = True lowercase__ : List[Any] = {'add_prefix_space': True} lowercase__ : str = False def snake_case__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] _lowerCamelCase = {'''unk_token''': '''<unk>'''} _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(lowerCamelCase__ ) ) def snake_case__ ( self , **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def snake_case__ ( self , **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = '''lower newer''' _lowerCamelCase = '''lower newer''' return input_text, output_text def snake_case__ ( self ): _lowerCamelCase = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _lowerCamelCase = '''lower newer''' _lowerCamelCase = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokens + [tokenizer.unk_token] _lowerCamelCase = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer(add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = '''lower newer''' # Testing tokenization _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids without special tokens _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids with special tokens _lowerCamelCase = self.get_rust_tokenizer(add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing the unknown token _lowerCamelCase = tokens + [rust_tokenizer.unk_token] _lowerCamelCase = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): # It's very difficult to mix/test pretokenization with byte-level # And get both GPT2 and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def snake_case__ ( self , lowerCamelCase__=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) # Simple input _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] _lowerCamelCase = ('''This is a simple input''', '''This is a pair''') _lowerCamelCase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(lowerCamelCase__ , tokenizer_r.encode , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(lowerCamelCase__ , tokenizer_r.encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( lowerCamelCase__ , tokenizer_r.batch_encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(lowerCamelCase__ , tokenizer_r.encode , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(lowerCamelCase__ , tokenizer_r.encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( lowerCamelCase__ , tokenizer_r.batch_encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' , ) def snake_case__ ( self ): _lowerCamelCase = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input looooooooong''', '''This is a simple input'''] _lowerCamelCase = ('''This is a simple input''', '''This is a pair''') _lowerCamelCase = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] _lowerCamelCase = tokenizer.pad_token_id _lowerCamelCase = tokenizer(lowerCamelCase__ , padding='''max_length''' , max_length=3_0 , return_tensors='''np''' ) _lowerCamelCase = tokenizer(lowerCamelCase__ , padding=lowerCamelCase__ , truncate=lowerCamelCase__ , return_tensors='''np''' ) _lowerCamelCase = tokenizer(*lowerCamelCase__ , padding='''max_length''' , max_length=6_0 , return_tensors='''np''' ) _lowerCamelCase = tokenizer(lowerCamelCase__ , padding=lowerCamelCase__ , truncate=lowerCamelCase__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def snake_case__ ( self ): _lowerCamelCase = '''$$$''' _lowerCamelCase = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=lowerCamelCase__ , add_bos_token=lowerCamelCase__ ) _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = tokenizer(lowerCamelCase__ ) _lowerCamelCase = tokenizer(lowerCamelCase__ ) self.assertEqual(out_s.input_ids[0] , lowerCamelCase__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _lowerCamelCase = tokenizer.decode(out_s.input_ids ) _lowerCamelCase = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , lowerCamelCase__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def snake_case__ ( self ): pass def snake_case__ ( self ): # TODO: change to self.get_tokenizers() when the fast version is implemented _lowerCamelCase = [self.get_tokenizer(do_lower_case=lowerCamelCase__ , add_bos_token=lowerCamelCase__ )] for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): _lowerCamelCase = '''Encode this.''' _lowerCamelCase = '''This one too please.''' _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) encoded_sequence += tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode_plus( lowerCamelCase__ , lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_special_tokens_mask=lowerCamelCase__ , ) _lowerCamelCase = encoded_sequence_dict['''input_ids'''] _lowerCamelCase = encoded_sequence_dict['''special_tokens_mask'''] self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) _lowerCamelCase = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(lowerCamelCase__ ) ] _lowerCamelCase = [x for x in filtered_sequence if x is not None] self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @require_tokenizers class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # More context: # https://huggingface.co/wjmcat/opt-350m-paddle/discussions/1 # https://huggingface.slack.com/archives/C01N44FJDHT/p1653511495183519 # https://github.com/huggingface/transformers/pull/17088#discussion_r871246439 _lowerCamelCase = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=lowerCamelCase__ ) _lowerCamelCase = '''A photo of a cat''' _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) self.assertEqual(lowerCamelCase__ , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('''test_opt''' ) _lowerCamelCase = AutoTokenizer.from_pretrained('''./test_opt''' ) _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) self.assertEqual(lowerCamelCase__ , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) def snake_case__ ( self ): _lowerCamelCase = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , use_slow=lowerCamelCase__ ) _lowerCamelCase = '''A photo of a cat''' _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) # Same as above self.assertEqual(lowerCamelCase__ , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) @unittest.skip('''This test is failing because of a bug in the fast tokenizer''' ) def snake_case__ ( self ): _lowerCamelCase = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=lowerCamelCase__ ) _lowerCamelCase = '''bos''' _lowerCamelCase = tokenizer.get_vocab()['''bos'''] _lowerCamelCase = '''A photo of a cat''' _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) # We changed the bos token self.assertEqual(lowerCamelCase__ , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('''./tok''' ) _lowerCamelCase = AutoTokenizer.from_pretrained('''./tok''' ) self.assertTrue(tokenizer.is_fast ) _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) self.assertEqual(lowerCamelCase__ , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] )
715
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
0
"""simple docstring""" import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel __SCREAMING_SNAKE_CASE : Optional[int] = '''0.12''' # assumed parallelism: 8 @require_flax @is_staging_test class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @classmethod def snake_case__ ( cls ): _lowerCamelCase = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def snake_case__ ( cls ): try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def snake_case__ ( self ): _lowerCamelCase = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) _lowerCamelCase = FlaxBertModel(lowerCamelCase__ ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) _lowerCamelCase = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) _lowerCamelCase = flatten_dict(unfreeze(model.params ) ) _lowerCamelCase = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _lowerCamelCase = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowerCamelCase__ , 1e-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ , repo_id='''test-model-flax''' , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _lowerCamelCase = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) _lowerCamelCase = flatten_dict(unfreeze(model.params ) ) _lowerCamelCase = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _lowerCamelCase = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowerCamelCase__ , 1e-3 , msg=F"""{key} not identical""" ) def snake_case__ ( self ): _lowerCamelCase = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) _lowerCamelCase = FlaxBertModel(lowerCamelCase__ ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) _lowerCamelCase = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) _lowerCamelCase = flatten_dict(unfreeze(model.params ) ) _lowerCamelCase = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _lowerCamelCase = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowerCamelCase__ , 1e-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( lowerCamelCase__ , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _lowerCamelCase = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) _lowerCamelCase = flatten_dict(unfreeze(model.params ) ) _lowerCamelCase = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _lowerCamelCase = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowerCamelCase__ , 1e-3 , msg=F"""{key} not identical""" ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : List[Any] ) -> Optional[Any]: _lowerCamelCase = True _lowerCamelCase = flatten_dict(modela.params ) _lowerCamelCase = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: _lowerCamelCase = False return models_are_equal @require_flax class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) _lowerCamelCase = FlaxBertModel(lowerCamelCase__ ) _lowerCamelCase = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(lowerCamelCase__ , lowerCamelCase__ ) ) with self.assertRaises(lowerCamelCase__ ): _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ , subfolder=lowerCamelCase__ ) self.assertTrue(check_models_equal(lowerCamelCase__ , lowerCamelCase__ ) ) def snake_case__ ( self ): _lowerCamelCase = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) _lowerCamelCase = FlaxBertModel(lowerCamelCase__ ) _lowerCamelCase = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(lowerCamelCase__ , lowerCamelCase__ ) , max_shard_size='''10KB''' ) with self.assertRaises(lowerCamelCase__ ): _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ , subfolder=lowerCamelCase__ ) self.assertTrue(check_models_equal(lowerCamelCase__ , lowerCamelCase__ ) ) def snake_case__ ( self ): _lowerCamelCase = '''bert''' _lowerCamelCase = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(lowerCamelCase__ ): _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ , subfolder=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = '''bert''' _lowerCamelCase = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(lowerCamelCase__ ): _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = FlaxBertModel.from_pretrained(lowerCamelCase__ , subfolder=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
716
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
623
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/config.json''', # See all XGLM models at https://huggingface.co/models?filter=xglm } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Tuple = 'xglm' lowercase__ : Tuple = ['past_key_values'] lowercase__ : Optional[int] = { 'num_attention_heads': 'attention_heads', 'hidden_size': 'd_model', 'num_hidden_layers': 'num_layers', } def __init__( self , lowerCamelCase__=2_5_6_0_0_8 , lowerCamelCase__=2_0_4_8 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=4_0_9_6 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , **lowerCamelCase__ , ): _lowerCamelCase = vocab_size _lowerCamelCase = max_position_embeddings _lowerCamelCase = d_model _lowerCamelCase = ffn_dim _lowerCamelCase = num_layers _lowerCamelCase = attention_heads _lowerCamelCase = activation_function _lowerCamelCase = dropout _lowerCamelCase = attention_dropout _lowerCamelCase = activation_dropout _lowerCamelCase = layerdrop _lowerCamelCase = init_std _lowerCamelCase = scale_embedding # scale factor will be sqrt(d_model) if True _lowerCamelCase = use_cache super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
717
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
0
"""simple docstring""" import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Union[str, Any]=False ) -> Dict: try: _lowerCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. _lowerCamelCase = default else: # KEY is set, convert it to True or False. try: _lowerCamelCase = strtobool(lowercase_ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F"""If set, {key} must be yes or no.""" ) return _value __SCREAMING_SNAKE_CASE : int = parse_flag_from_env('''RUN_SLOW''', default=False) def lowerCAmelCase_( lowercase_ : str ) -> Union[str, Any]: return unittest.skip('''Test was skipped''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> Tuple: return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Dict ) -> Optional[int]: return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Optional[int]: return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Tuple ) -> int: return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[Any] ) -> List[str]: return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> str: return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> Dict: return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Any ) -> List[str]: return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> Dict: return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[Any] ) -> List[Any]: return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Optional[Any]: return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Any ) -> Tuple: return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> List[Any]: return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : str ) -> List[Any]: return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : str ) -> int: return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[Any]=None , lowercase_ : List[Any]=None ) -> List[Any]: if test_case is None: return partial(lowercase_ , version=lowercase_ ) return unittest.skipUnless(is_torch_version('''>=''' , lowercase_ ) , F"""test requires torch version >= {version}""" )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[int]: return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> str: return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowercase_ ) def lowerCAmelCase_( lowercase_ : Tuple ) -> List[Any]: return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowercase_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Tuple: return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowercase_ ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = True @classmethod def snake_case__ ( cls ): _lowerCamelCase = tempfile.mkdtemp() @classmethod def snake_case__ ( cls ): if os.path.exists(cls.tmpdir ): shutil.rmtree(cls.tmpdir ) def snake_case__ ( self ): if self.clear_on_setup: for path in Path(self.tmpdir ).glob('''**/*''' ): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(lowerCamelCase__ ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = mocks if isinstance(lowerCamelCase__ , (tuple, list) ) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Any: _lowerCamelCase = AcceleratorState() _lowerCamelCase = tensor[None].clone().to(state.device ) _lowerCamelCase = gather(lowercase_ ).cpu() _lowerCamelCase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowercase_ ): return False return True class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = returncode _lowerCamelCase = stdout _lowerCamelCase = stderr async def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : int ) -> Optional[int]: while True: _lowerCamelCase = await stream.readline() if line: callback(lowercase_ ) else: break async def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : Union[str, Any]=None , lowercase_ : Tuple=None , lowercase_ : List[str]=None , lowercase_ : Union[str, Any]=False , lowercase_ : Union[str, Any]=False ) -> _RunOutput: if echo: print('''\nRunning: ''' , ''' '''.join(lowercase_ ) ) _lowerCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowercase_ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowercase_ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) _lowerCamelCase = [] _lowerCamelCase = [] def tee(lowercase_ : int , lowercase_ : List[Any] , lowercase_ : Dict , lowercase_ : List[str]="" ): _lowerCamelCase = line.decode('''utf-8''' ).rstrip() sink.append(lowercase_ ) if not quiet: print(lowercase_ , lowercase_ , file=lowercase_ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowercase_ : tee(lowercase_ , lowercase_ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowercase_ : tee(lowercase_ , lowercase_ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowercase_ , ) return _RunOutput(await p.wait() , lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Optional[Any]=None , lowercase_ : Any=None , lowercase_ : Any=1_80 , lowercase_ : Optional[int]=False , lowercase_ : Optional[Any]=True ) -> _RunOutput: _lowerCamelCase = asyncio.get_event_loop() _lowerCamelCase = loop.run_until_complete( _stream_subprocess(lowercase_ , env=lowercase_ , stdin=lowercase_ , timeout=lowercase_ , quiet=lowercase_ , echo=lowercase_ ) ) _lowerCamelCase = ''' '''.join(lowercase_ ) if result.returncode > 0: _lowerCamelCase = '''\n'''.join(result.stderr ) raise RuntimeError( F"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" F"""The combined stderr from workers follows:\n{stderr}""" ) return result class lowerCamelCase_( A__ ): '''simple docstring''' pass def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Dict=False ) -> Optional[int]: try: _lowerCamelCase = subprocess.check_output(lowercase_ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowercase_ , '''decode''' ): _lowerCamelCase = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F"""Command `{" ".join(lowercase_ )}` failed with the following error:\n\n{e.output.decode()}""" ) from e
718
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
0
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int , lowercase_ : float , lowercase_ : float ) -> float: return round(float(moles / volume ) * nfactor ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: return round(float((moles * 0.0_8_2_1 * temperature) / (volume) ) ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: return round(float((moles * 0.0_8_2_1 * temperature) / (pressure) ) ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: return round(float((pressure * volume) / (0.0_8_2_1 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
719
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
0
"""simple docstring""" from scipy.stats import spearmanr import datasets __SCREAMING_SNAKE_CASE : Optional[Any] = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __SCREAMING_SNAKE_CASE : Tuple = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __SCREAMING_SNAKE_CASE : Dict = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = spearmanr(lowerCamelCase__ , lowerCamelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
720
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
0
"""simple docstring""" import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py __SCREAMING_SNAKE_CASE : Any = '''src/transformers''' __SCREAMING_SNAKE_CASE : Optional[int] = '''docs/source/en/tasks''' def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : List[Any] ) -> Tuple: with open(lowercase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: _lowerCamelCase = f.readlines() # Find the start prompt. _lowerCamelCase = 0 while not lines[start_index].startswith(lowercase_ ): start_index += 1 start_index += 1 _lowerCamelCase = start_index while not lines[end_index].startswith(lowercase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. __SCREAMING_SNAKE_CASE : Optional[Any] = direct_transformers_import(TRANSFORMERS_PATH) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). __SCREAMING_SNAKE_CASE : List[str] = { '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> List[Any]: _lowerCamelCase = TASK_GUIDE_TO_MODELS[task_guide] _lowerCamelCase = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(lowercase_ , set() ) _lowerCamelCase = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"""[{name}](../model_doc/{code})""" for code, name in model_names.items()] ) + "\n" def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Optional[int]=False ) -> str: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = _find_text_in_file( filename=os.path.join(lowercase_ , lowercase_ ) , start_prompt='''<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->''' , end_prompt='''<!--End of the generated tip-->''' , ) _lowerCamelCase = get_model_list_for_task(lowercase_ ) if current_list != new_list: if overwrite: with open(os.path.join(lowercase_ , lowercase_ ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"""The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`""" ''' to fix this.''' ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') __SCREAMING_SNAKE_CASE : List[str] = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
721
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
"""simple docstring""" import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration __SCREAMING_SNAKE_CASE : List[str] = 5_0_0_0_0_0 __SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.split(__file__) __SCREAMING_SNAKE_CASE : Any = os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def lowerCAmelCase_( lowercase_ : datasets.Dataset , **lowercase_ : Union[str, Any] ) -> List[str]: _lowerCamelCase = dataset.map(**lowercase_ ) @get_duration def lowerCAmelCase_( lowercase_ : datasets.Dataset , **lowercase_ : Optional[int] ) -> Any: _lowerCamelCase = dataset.filter(**lowercase_ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = {'''num examples''': SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase = datasets.Features({'''text''': datasets.Value('''string''' ), '''numbers''': datasets.Value('''float32''' )} ) _lowerCamelCase = generate_example_dataset( os.path.join(lowercase_ , '''dataset.arrow''' ) , lowercase_ , num_examples=lowercase_ ) _lowerCamelCase = transformers.AutoTokenizer.from_pretrained('''bert-base-cased''' , use_fast=lowercase_ ) def tokenize(lowercase_ : Dict ): return tokenizer(examples['''text'''] ) _lowerCamelCase = map(lowercase_ ) _lowerCamelCase = map(lowercase_ , batched=lowercase_ ) _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''numpy''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''pandas''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''torch''' , columns='''numbers''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''tensorflow''' , columns='''numbers''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) _lowerCamelCase = map(lowercase_ , function=lowercase_ , batched=lowercase_ ) _lowerCamelCase = filter(lowercase_ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowercase_ , '''wb''' ) as f: f.write(json.dumps(lowercase_ ).encode('''utf-8''' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
700
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
0
"""simple docstring""" from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCamelCase_( A__, A__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None ): super().__init__() _lowerCamelCase = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" _lowerCamelCase = torch.zeros(lowerCamelCase__ , lowerCamelCase__ ) else: _lowerCamelCase = None _lowerCamelCase = torch.nn.Parameter(lowerCamelCase__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : VQModel lowercase__ : CLIPTextModel lowercase__ : CLIPTokenizer lowercase__ : TransformeraDModel lowercase__ : LearnedClassifierFreeSamplingEmbeddings lowercase__ : VQDiffusionScheduler def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): super().__init__() self.register_modules( vqvae=lowerCamelCase__ , transformer=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , scheduler=lowerCamelCase__ , learned_classifier_free_sampling_embeddings=lowerCamelCase__ , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = len(lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else 1 # get prompt text embeddings _lowerCamelCase = self.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=self.tokenizer.model_max_length , return_tensors='''pt''' , ) _lowerCamelCase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: _lowerCamelCase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) _lowerCamelCase = text_input_ids[:, : self.tokenizer.model_max_length] _lowerCamelCase = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 _lowerCamelCase = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=lowerCamelCase__ ) # duplicate text embeddings for each generation per prompt _lowerCamelCase = prompt_embeds.repeat_interleave(lowerCamelCase__ , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: _lowerCamelCase = self.learned_classifier_free_sampling_embeddings.embeddings _lowerCamelCase = negative_prompt_embeds.unsqueeze(0 ).repeat(lowerCamelCase__ , 1 , 1 ) else: _lowerCamelCase = [''''''] * batch_size _lowerCamelCase = text_input_ids.shape[-1] _lowerCamelCase = self.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=lowerCamelCase__ , truncation=lowerCamelCase__ , return_tensors='''pt''' , ) _lowerCamelCase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings _lowerCamelCase = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=lowerCamelCase__ ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _lowerCamelCase = negative_prompt_embeds.shape[1] _lowerCamelCase = negative_prompt_embeds.repeat(1 , lowerCamelCase__ , 1 ) _lowerCamelCase = negative_prompt_embeds.view(batch_size * num_images_per_prompt , lowerCamelCase__ , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _lowerCamelCase = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , lowerCamelCase__ , lowerCamelCase__ = 1_0_0 , lowerCamelCase__ = 5.0 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = 1 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = "pil" , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = 1 , ): if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = 1 elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = len(lowerCamelCase__ ) else: raise ValueError(F"""`prompt` has to be of type `str` or `list` but is {type(lowerCamelCase__ )}""" ) _lowerCamelCase = batch_size * num_images_per_prompt _lowerCamelCase = guidance_scale > 1.0 _lowerCamelCase = self._encode_prompt(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(lowerCamelCase__ , lowerCamelCase__ ) or callback_steps <= 0) ): raise ValueError( F"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" F""" {type(lowerCamelCase__ )}.""" ) # get the initial completely masked latents unless the user supplied it _lowerCamelCase = (batch_size, self.transformer.num_latent_pixels) if latents is None: _lowerCamelCase = self.transformer.num_vector_embeds - 1 _lowerCamelCase = torch.full(lowerCamelCase__ , lowerCamelCase__ ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( '''Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,''' F""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) _lowerCamelCase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) _lowerCamelCase = self.scheduler.timesteps.to(self.device ) _lowerCamelCase = latents for i, t in enumerate(self.progress_bar(lowerCamelCase__ ) ): # expand the sample if we are doing classifier free guidance _lowerCamelCase = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` _lowerCamelCase = self.transformer(lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , timestep=lowerCamelCase__ ).sample if do_classifier_free_guidance: _lowerCamelCase , _lowerCamelCase = model_output.chunk(2 ) _lowerCamelCase = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(lowerCamelCase__ , dim=1 , keepdim=lowerCamelCase__ ) _lowerCamelCase = self.truncate(lowerCamelCase__ , lowerCamelCase__ ) # remove `log(0)`'s (`-inf`s) _lowerCamelCase = model_output.clamp(-7_0 ) # compute the previous noisy sample x_t -> x_t-1 _lowerCamelCase = self.scheduler.step(lowerCamelCase__ , timestep=lowerCamelCase__ , sample=lowerCamelCase__ , generator=lowerCamelCase__ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.vqvae.config.vq_embed_dim _lowerCamelCase = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) _lowerCamelCase = self.vqvae.quantize.get_codebook_entry(lowerCamelCase__ , shape=lowerCamelCase__ ) _lowerCamelCase = self.vqvae.decode(lowerCamelCase__ , force_not_quantize=lowerCamelCase__ ).sample _lowerCamelCase = (image / 2 + 0.5).clamp(0 , 1 ) _lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _lowerCamelCase = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase , _lowerCamelCase = torch.sort(lowerCamelCase__ , 1 , descending=lowerCamelCase__ ) _lowerCamelCase = torch.exp(lowerCamelCase__ ) _lowerCamelCase = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out _lowerCamelCase = torch.full_like(keep_mask[:, 0:1, :] , lowerCamelCase__ ) _lowerCamelCase = torch.cat((all_true, keep_mask) , dim=1 ) _lowerCamelCase = keep_mask[:, :-1, :] _lowerCamelCase = keep_mask.gather(1 , indices.argsort(1 ) ) _lowerCamelCase = log_p_x_0.clone() _lowerCamelCase = -torch.inf # -inf = log(0) return rv
701
"""simple docstring""" from typing import Any class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) _lowerCamelCase = temp.next print() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = Node(lowerCamelCase__ ) _lowerCamelCase = self.head _lowerCamelCase = new_node def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if node_data_a == node_data_a: return else: _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next if node_a is None or node_a is None: return _lowerCamelCase , _lowerCamelCase = node_a.data, node_a.data if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
623
0
"""simple docstring""" import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer __SCREAMING_SNAKE_CASE : int = '''bart''' __SCREAMING_SNAKE_CASE : int = True @st.cache(allow_output_mutation=lowercase_ ) def lowerCAmelCase_( ) -> Tuple: if LOAD_DENSE_INDEX: _lowerCamelCase = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) _lowerCamelCase = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) _lowerCamelCase = qar_model.eval() else: _lowerCamelCase , _lowerCamelCase = (None, None) if MODEL_TYPE == "bart": _lowerCamelCase = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) _lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) _lowerCamelCase = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) _lowerCamelCase = sas_model.eval() else: _lowerCamelCase , _lowerCamelCase = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=lowercase_ ) def lowerCAmelCase_( ) -> Union[str, Any]: if LOAD_DENSE_INDEX: _lowerCamelCase = faiss.StandardGpuResources() _lowerCamelCase = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] _lowerCamelCase = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 1_28) , ) _lowerCamelCase = faiss.IndexFlatIP(1_28 ) _lowerCamelCase = faiss.index_cpu_to_gpu(lowercase_ , 1 , lowercase_ ) wikiaab_gpu_index_flat.add(lowercase_ ) # TODO fix for larger GPU else: _lowerCamelCase , _lowerCamelCase = (None, None) _lowerCamelCase = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=lowercase_ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) _lowerCamelCase = elia['''train_eli5'''] _lowerCamelCase = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 1_28) ) _lowerCamelCase = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(lowercase_ ) return (elia_train, eli5_train_q_index) __SCREAMING_SNAKE_CASE : Dict = load_indexes() __SCREAMING_SNAKE_CASE : Any = load_models() __SCREAMING_SNAKE_CASE : str = load_train_data() def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Any=10 ) -> int: _lowerCamelCase = embed_questions_for_retrieval([question] , lowercase_ , lowercase_ ) _lowerCamelCase , _lowerCamelCase = eli5_train_q_index.search(lowercase_ , lowercase_ ) _lowerCamelCase = [elia_train[int(lowercase_ )] for i in I[0]] return nn_examples def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Optional[Any]="wiki40b" , lowercase_ : Dict="dense" , lowercase_ : Dict=10 ) -> int: if source == "none": _lowerCamelCase , _lowerCamelCase = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": _lowerCamelCase , _lowerCamelCase = query_qa_dense_index( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: _lowerCamelCase , _lowerCamelCase = query_es_index( lowercase_ , lowercase_ , index_name='''english_wiki40b_snippets_100w''' , n_results=lowercase_ , ) _lowerCamelCase = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] _lowerCamelCase = '''question: {} context: {}'''.format(lowercase_ , lowercase_ ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda lowercase_ : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda lowercase_ : None), } ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Any=64 , lowercase_ : Union[str, Any]=2_56 , lowercase_ : List[str]=False , lowercase_ : Optional[int]=2 , lowercase_ : Union[str, Any]=0.9_5 , lowercase_ : List[str]=0.8 ) -> Dict: with torch.no_grad(): _lowerCamelCase = qa_sas_generate( lowercase_ , lowercase_ , lowercase_ , num_answers=1 , num_beams=lowercase_ , min_len=lowercase_ , max_len=lowercase_ , do_sample=lowercase_ , temp=lowercase_ , top_p=lowercase_ , top_k=lowercase_ , max_input_length=10_24 , device='''cuda:0''' , )[0] return (answer, support_list) st.title('''Long Form Question Answering with ELI5''') # Start sidebar __SCREAMING_SNAKE_CASE : Any = '''<img src=\'https://huggingface.co/front/assets/huggingface_logo.svg\'>''' __SCREAMING_SNAKE_CASE : int = ''' <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class="img-container"> <!-- Inline parent element --> %s </span> </body> </html> ''' % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia __SCREAMING_SNAKE_CASE : int = ''' This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. ''' st.sidebar.markdown(description, unsafe_allow_html=True) __SCREAMING_SNAKE_CASE : List[str] = [ '''Answer the question''', '''View the retrieved document only''', '''View the most similar ELI5 question and answer''', '''Show me everything, please!''', ] __SCREAMING_SNAKE_CASE : Optional[int] = st.sidebar.checkbox('''Demo options''') if demo_options: __SCREAMING_SNAKE_CASE : Dict = st.sidebar.selectbox( '''''', action_list, index=3, ) __SCREAMING_SNAKE_CASE : Any = action_list.index(action_st) __SCREAMING_SNAKE_CASE : str = st.sidebar.selectbox( '''''', ['''Show full text of passages''', '''Show passage section titles'''], index=0, ) __SCREAMING_SNAKE_CASE : Optional[int] = show_type == '''Show full text of passages''' else: __SCREAMING_SNAKE_CASE : Any = 3 __SCREAMING_SNAKE_CASE : Union[str, Any] = True __SCREAMING_SNAKE_CASE : Tuple = st.sidebar.checkbox('''Retrieval options''') if retrieval_options: __SCREAMING_SNAKE_CASE : Optional[int] = ''' ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. ''' st.sidebar.markdown(retriever_info) __SCREAMING_SNAKE_CASE : Any = st.sidebar.selectbox('''Which Wikipedia format should the model use?''', ['''wiki40b''', '''none''']) __SCREAMING_SNAKE_CASE : Optional[Any] = st.sidebar.selectbox('''Which Wikipedia indexer should the model use?''', ['''dense''', '''sparse''', '''mixed''']) else: __SCREAMING_SNAKE_CASE : Optional[Any] = '''wiki40b''' __SCREAMING_SNAKE_CASE : Dict = '''dense''' __SCREAMING_SNAKE_CASE : Any = '''beam''' __SCREAMING_SNAKE_CASE : int = 2 __SCREAMING_SNAKE_CASE : Dict = 6_4 __SCREAMING_SNAKE_CASE : int = 2_5_6 __SCREAMING_SNAKE_CASE : Tuple = None __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : Tuple = st.sidebar.checkbox('''Generation options''') if generate_options: __SCREAMING_SNAKE_CASE : int = ''' ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder\'s output probabilities. ''' st.sidebar.markdown(generate_info) __SCREAMING_SNAKE_CASE : Union[str, Any] = st.sidebar.selectbox('''Would you like to use beam search or sample an answer?''', ['''beam''', '''sampled''']) __SCREAMING_SNAKE_CASE : Optional[Any] = st.sidebar.slider( '''Minimum generation length''', min_value=8, max_value=2_5_6, value=6_4, step=8, format=None, key=None ) __SCREAMING_SNAKE_CASE : Optional[int] = st.sidebar.slider( '''Maximum generation length''', min_value=6_4, max_value=5_1_2, value=2_5_6, step=1_6, format=None, key=None ) if sampled == "beam": __SCREAMING_SNAKE_CASE : str = st.sidebar.slider('''Beam size''', min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: __SCREAMING_SNAKE_CASE : str = st.sidebar.slider( '''Nucleus sampling p''', min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) __SCREAMING_SNAKE_CASE : Optional[Any] = st.sidebar.slider( '''Temperature''', min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) __SCREAMING_SNAKE_CASE : Union[str, Any] = None # start main text __SCREAMING_SNAKE_CASE : Optional[Any] = [ '''<MY QUESTION>''', '''How do people make chocolate?''', '''Why do we get a fever when we are sick?''', '''How can different animals perceive different colors?''', '''What is natural language processing?''', '''What\'s the best way to treat a sunburn?''', '''What exactly are vitamins ?''', '''How does nuclear energy provide electricity?''', '''What\'s the difference between viruses and bacteria?''', '''Why are flutes classified as woodwinds when most of them are made out of metal ?''', '''Why do people like drinking coffee even though it tastes so bad?''', '''What happens when wine ages? How does it make the wine taste better?''', '''If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?''', '''How can we set a date to the beginning or end of an artistic period? Doesn\'t the change happen gradually?''', '''How does New Zealand have so many large bird predators?''', ] __SCREAMING_SNAKE_CASE : int = st.selectbox( '''What would you like to ask? ---- select <MY QUESTION> to enter a new query''', questions_list, index=1, ) if question_s == "<MY QUESTION>": __SCREAMING_SNAKE_CASE : Optional[int] = st.text_input('''Enter your question here:''', '''''') else: __SCREAMING_SNAKE_CASE : Optional[int] = question_s if st.button('''Show me!'''): if action in [0, 1, 3]: if index_type == "mixed": __SCREAMING_SNAKE_CASE : Any = make_support(question, source=wiki_source, method='''dense''', n_results=1_0) __SCREAMING_SNAKE_CASE : List[str] = make_support(question, source=wiki_source, method='''sparse''', n_results=1_0) __SCREAMING_SNAKE_CASE : List[str] = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] __SCREAMING_SNAKE_CASE : Any = support_list[:1_0] __SCREAMING_SNAKE_CASE : Optional[int] = '''<P> ''' + ''' <P> '''.join([res[-1] for res in support_list]) else: __SCREAMING_SNAKE_CASE : Optional[Any] = make_support(question, source=wiki_source, method=index_type, n_results=1_0) if action in [0, 3]: __SCREAMING_SNAKE_CASE : Dict = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == '''sampled'''), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown('''### The model generated answer is:''') st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown('''--- \n ### The model is drawing information from the following Wikipedia passages:''') for i, res in enumerate(support_list): __SCREAMING_SNAKE_CASE : int = '''https://en.wikipedia.org/wiki/{}'''.format(res[0].replace(''' ''', '''_''')) __SCREAMING_SNAKE_CASE : Any = res[1].strip() if sec_titles == "": __SCREAMING_SNAKE_CASE : Dict = '''[{}]({})'''.format(res[0], wiki_url) else: __SCREAMING_SNAKE_CASE : List[str] = sec_titles.split(''' & ''') __SCREAMING_SNAKE_CASE : List[str] = ''' & '''.join( ['''[{}]({}#{})'''.format(sec.strip(), wiki_url, sec.strip().replace(''' ''', '''_''')) for sec in sec_list] ) st.markdown( '''{0:02d} - **Article**: {1:<18} <br> _Section_: {2}'''.format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( '''> <span style="font-family:arial; font-size:10pt;">''' + res[-1] + '''</span>''', unsafe_allow_html=True ) if action in [2, 3]: __SCREAMING_SNAKE_CASE : Dict = find_nearest_training(question) __SCREAMING_SNAKE_CASE : Optional[Any] = nn_train_list[0] st.markdown( '''--- \n ### The most similar question in the ELI5 training set was: \n\n {}'''.format(train_exple['''title''']) ) __SCREAMING_SNAKE_CASE : Optional[int] = [ '''{}. {}'''.format(i + 1, ''' \n'''.join([line.strip() for line in ans.split('''\n''') if line.strip() != ''''''])) for i, (ans, sc) in enumerate(zip(train_exple['''answers''']['''text'''], train_exple['''answers''']['''score'''])) if i == 0 or sc > 2 ] st.markdown('''##### Its answers were: \n\n {}'''.format('''\n'''.join(answers_st))) __SCREAMING_SNAKE_CASE : str = ''' --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* ''' st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
702
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __SCREAMING_SNAKE_CASE : Optional[Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[Any]: from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> List[str]: from diffusers.utils.testing_utils import pytest_terminal_summary_main _lowerCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
623
0
"""simple docstring""" import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_ckpt''' , type=lowercase_ , default='''microsoft/unixcoder-base-nine''' ) parser.add_argument('''--num_epochs''' , type=lowercase_ , default=5 ) parser.add_argument('''--batch_size''' , type=lowercase_ , default=6 ) parser.add_argument('''--gradient_accumulation_steps''' , type=lowercase_ , default=1 ) parser.add_argument('''--freeze''' , type=lowercase_ , default=lowercase_ ) parser.add_argument('''--learning_rate''' , type=lowercase_ , default=5e-4 ) parser.add_argument('''--seed''' , type=lowercase_ , default=0 ) parser.add_argument('''--lr_scheduler_type''' , type=lowercase_ , default='''cosine''' ) parser.add_argument('''--num_warmup_steps''' , type=lowercase_ , default=10 ) parser.add_argument('''--weight_decay''' , type=lowercase_ , default=0.0_1 ) parser.add_argument('''--output_dir''' , type=lowercase_ , default='''./results''' ) return parser.parse_args() __SCREAMING_SNAKE_CASE : Any = load('''accuracy''') def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase , _lowerCamelCase = eval_pred _lowerCamelCase = np.argmax(lowercase_ , axis=1 ) return metric.compute(predictions=lowercase_ , references=lowercase_ ) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__() _lowerCamelCase = trainer def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ): if control.should_evaluate: _lowerCamelCase = deepcopy(lowerCamelCase__ ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix='''train''' ) return control_copy def lowerCAmelCase_( ) -> Any: _lowerCamelCase = get_args() set_seed(args.seed ) _lowerCamelCase = load_dataset('''codeparrot/codecomplex''' , split='''train''' ) _lowerCamelCase = dataset.train_test_split(test_size=0.2 ) _lowerCamelCase = train_test['''test'''].train_test_split(test_size=0.5 ) _lowerCamelCase = DatasetDict( { '''train''': train_test['''train'''], '''test''': test_validation['''train'''], '''valid''': test_validation['''test'''], } ) print('''Loading tokenizer and model''' ) _lowerCamelCase = AutoTokenizer.from_pretrained(args.model_ckpt ) _lowerCamelCase = tokenizer.eos_token _lowerCamelCase = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt , num_labels=7 ) _lowerCamelCase = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): _lowerCamelCase = False _lowerCamelCase = ClassLabel(num_classes=7 , names=list(set(train_test_validation['''train''']['''complexity'''] ) ) ) def tokenize(lowercase_ : Union[str, Any] ): _lowerCamelCase = tokenizer(example['''src'''] , truncation=lowercase_ , max_length=10_24 ) _lowerCamelCase = labels.straint(example['''complexity'''] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } _lowerCamelCase = train_test_validation.map( lowercase_ , batched=lowercase_ , remove_columns=train_test_validation['''train'''].column_names , ) _lowerCamelCase = DataCollatorWithPadding(tokenizer=lowercase_ ) _lowerCamelCase = TrainingArguments( output_dir=args.output_dir , learning_rate=args.learning_rate , lr_scheduler_type=args.lr_scheduler_type , evaluation_strategy='''epoch''' , save_strategy='''epoch''' , logging_strategy='''epoch''' , per_device_train_batch_size=args.batch_size , per_device_eval_batch_size=args.batch_size , num_train_epochs=args.num_epochs , gradient_accumulation_steps=args.gradient_accumulation_steps , weight_decay=0.0_1 , metric_for_best_model='''accuracy''' , run_name='''complexity-java''' , report_to='''wandb''' , ) _lowerCamelCase = Trainer( model=lowercase_ , args=lowercase_ , train_dataset=tokenized_datasets['''train'''] , eval_dataset=tokenized_datasets['''valid'''] , tokenizer=lowercase_ , data_collator=lowercase_ , compute_metrics=lowercase_ , ) print('''Training...''' ) trainer.add_callback(CustomCallback(lowercase_ ) ) trainer.train() if __name__ == "__main__": main()
703
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
704
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
0
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __SCREAMING_SNAKE_CASE : int = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.14.0''', '''To fix: pip install -r examples/pytorch/audio-classification/requirements.txt''') def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : float , lowercase_ : int = 1_60_00 ) -> Tuple: _lowerCamelCase = int(round(sample_rate * max_length ) ) if len(lowercase_ ) <= sample_length: return wav _lowerCamelCase = randint(0 , len(lowercase_ ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : Optional[str] = field(default=A__, metadata={'help': 'Name of a dataset from the datasets package'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'A file containing the training audio paths and labels.'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'A file containing the validation audio paths and labels.'} ) lowercase__ : str = field( default='train', metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' }, ) lowercase__ : str = field( default='validation', metadata={ 'help': ( 'The name of the training data set split to use (via the datasets library). Defaults to \'validation\'' ) }, ) lowercase__ : str = field( default='audio', metadata={'help': 'The name of the dataset column containing the audio data. Defaults to \'audio\''}, ) lowercase__ : str = field( default='label', metadata={'help': 'The name of the dataset column containing the labels. Defaults to \'label\''} ) lowercase__ : Optional[int] = field( default=A__, metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) }, ) lowercase__ : Optional[int] = field( default=A__, metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) }, ) lowercase__ : float = field( default=20, metadata={'help': 'Audio clips will be randomly cut to this length during training if the value is set.'}, ) @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : str = field( default='facebook/wav2vec2-base', metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'}, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Where do you want to store the pretrained models downloaded from the Hub'} ) lowercase__ : str = field( default='main', metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'}, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Name or path of preprocessor config.'} ) lowercase__ : bool = field( default=A__, metadata={'help': 'Whether to freeze the feature encoder layers of the model.'} ) lowercase__ : bool = field( default=A__, metadata={'help': 'Whether to generate an attention mask in the feature extractor.'} ) lowercase__ : bool = field( default=A__, metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) }, ) lowercase__ : Optional[bool] = field( default=A__, metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) lowercase__ : bool = field( default=A__, metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'}, ) def snake_case__ ( self ): if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( '''The argument `--freeze_feature_extractor` is deprecated and ''' '''will be removed in a future version. Use `--freeze_feature_encoder`''' '''instead. Setting `freeze_feature_encoder==True`.''' , lowerCamelCase__ , ) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( '''The argument `--freeze_feature_extractor` is deprecated and ''' '''should not be used in combination with `--freeze_feature_encoder`.''' '''Only make use of `--freeze_feature_encoder`.''' ) def lowerCAmelCase_( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowerCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_audio_classification''' , lowercase_ , lowercase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowerCamelCase = training_args.get_process_log_level() logger.setLevel(lowercase_ ) transformers.utils.logging.set_verbosity(lowercase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} """ + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. _lowerCamelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCamelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to train from scratch.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset and prepare it for the audio classification task. _lowerCamelCase = DatasetDict() _lowerCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( F"""--audio_column_name {data_args.audio_column_name} not found in dataset '{data_args.dataset_name}'. """ '''Make sure to set `--audio_column_name` to the correct audio column - one of ''' F"""{", ".join(raw_datasets["train"].column_names )}.""" ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( F"""--label_column_name {data_args.label_column_name} not found in dataset '{data_args.dataset_name}'. """ '''Make sure to set `--label_column_name` to the correct text column - one of ''' F"""{", ".join(raw_datasets["train"].column_names )}.""" ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy _lowerCamelCase = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. _lowerCamelCase = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) _lowerCamelCase = feature_extractor.model_input_names[0] def train_transforms(lowercase_ : Dict ): _lowerCamelCase = [] for audio in batch[data_args.audio_column_name]: _lowerCamelCase = random_subsample( audio['''array'''] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(lowercase_ ) _lowerCamelCase = feature_extractor(lowercase_ , sampling_rate=feature_extractor.sampling_rate ) _lowerCamelCase = {model_input_name: inputs.get(lowercase_ )} _lowerCamelCase = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(lowercase_ : str ): _lowerCamelCase = [audio['''array'''] for audio in batch[data_args.audio_column_name]] _lowerCamelCase = feature_extractor(lowercase_ , sampling_rate=feature_extractor.sampling_rate ) _lowerCamelCase = {model_input_name: inputs.get(lowercase_ )} _lowerCamelCase = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _lowerCamelCase = raw_datasets['''train'''].features[data_args.label_column_name].names _lowerCamelCase , _lowerCamelCase = {}, {} for i, label in enumerate(lowercase_ ): _lowerCamelCase = str(lowercase_ ) _lowerCamelCase = label # Load the accuracy metric from the datasets package _lowerCamelCase = evaluate.load('''accuracy''' ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(lowercase_ : Optional[int] ): _lowerCamelCase = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=lowercase_ , references=eval_pred.label_ids ) _lowerCamelCase = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(lowercase_ ) , labelaid=lowercase_ , idalabel=lowercase_ , finetuning_task='''audio-classification''' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowercase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: _lowerCamelCase = ( raw_datasets['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(lowercase_ , output_all_columns=lowercase_ ) if training_args.do_eval: if data_args.max_eval_samples is not None: _lowerCamelCase = ( raw_datasets['''eval'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(lowercase_ , output_all_columns=lowercase_ ) # Initialize our trainer _lowerCamelCase = Trainer( model=lowercase_ , args=lowercase_ , train_dataset=raw_datasets['''train'''] if training_args.do_train else None , eval_dataset=raw_datasets['''eval'''] if training_args.do_eval else None , compute_metrics=lowercase_ , tokenizer=lowercase_ , ) # Training if training_args.do_train: _lowerCamelCase = None if training_args.resume_from_checkpoint is not None: _lowerCamelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCamelCase = last_checkpoint _lowerCamelCase = trainer.train(resume_from_checkpoint=lowercase_ ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _lowerCamelCase = trainer.evaluate() trainer.log_metrics('''eval''' , lowercase_ ) trainer.save_metrics('''eval''' , lowercase_ ) # Write model card and (optionally) push to hub _lowerCamelCase = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''audio-classification''', '''dataset''': data_args.dataset_name, '''tags''': ['''audio-classification'''], } if training_args.push_to_hub: trainer.push_to_hub(**lowercase_ ) else: trainer.create_model_card(**lowercase_ ) if __name__ == "__main__": main()
705
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
706
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
0
"""simple docstring""" import itertools import math def lowerCAmelCase_( lowercase_ : int ) -> bool: '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowercase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCAmelCase_( ) -> List[str]: '''simple docstring''' _lowerCamelCase = 2 while True: if is_prime(lowercase_ ): yield num num += 1 def lowerCAmelCase_( lowercase_ : int = 1_00_01 ) -> int: '''simple docstring''' return next(itertools.islice(prime_generator() , nth - 1 , lowercase_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
707
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
0
"""simple docstring""" import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase_: def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=1_6 , lowerCamelCase__=[1, 2, 1] , lowerCamelCase__=[2, 2, 4] , lowerCamelCase__=2 , lowerCamelCase__=2.0 , lowerCamelCase__=True , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.1 , lowerCamelCase__="gelu" , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-5 , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=1_0 , lowerCamelCase__=8 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = embed_dim _lowerCamelCase = depths _lowerCamelCase = num_heads _lowerCamelCase = window_size _lowerCamelCase = mlp_ratio _lowerCamelCase = qkv_bias _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = drop_path_rate _lowerCamelCase = hidden_act _lowerCamelCase = use_absolute_embeddings _lowerCamelCase = patch_norm _lowerCamelCase = layer_norm_eps _lowerCamelCase = initializer_range _lowerCamelCase = is_training _lowerCamelCase = scope _lowerCamelCase = use_labels _lowerCamelCase = type_sequence_label_size _lowerCamelCase = encoder_stride def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = SwinvaModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _lowerCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = SwinvaForMaskedImageModeling(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = SwinvaForMaskedImageModeling(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.type_sequence_label_size _lowerCamelCase = SwinvaForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): lowercase__ : Tuple = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase__ : Optional[int] = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase__ : Any = False lowercase__ : Tuple = False lowercase__ : Tuple = False lowercase__ : Dict = False def snake_case__ ( self ): _lowerCamelCase = SwinvaModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , embed_dim=3_7 ) def snake_case__ ( self ): self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @unittest.skip(reason='''Got `CUDA error: misaligned address` with PyTorch 2.0.0.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Swinv2 does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = True for model_class in self.all_model_classes: _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.attentions _lowerCamelCase = len(self.model_tester.depths ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _lowerCamelCase = True _lowerCamelCase = config.window_size**2 _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.attentions self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) _lowerCamelCase = len(lowerCamelCase__ ) # Check attention is always last and order is fine _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) if hasattr(self.model_tester , '''num_hidden_states_types''' ): _lowerCamelCase = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states _lowerCamelCase = 2 self.assertEqual(out_len + added_hidden_states , len(lowerCamelCase__ ) ) _lowerCamelCase = outputs.attentions self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.hidden_states _lowerCamelCase = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # Swinv2 has a different seq_length _lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _lowerCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) _lowerCamelCase = outputs.reshaped_hidden_states self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = reshaped_hidden_states[0].shape _lowerCamelCase = ( reshaped_hidden_states[0].view(lowerCamelCase__ , lowerCamelCase__ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = 3 _lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _lowerCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _lowerCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = SwinvaModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = _config_zero_init(lowerCamelCase__ ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class lowerCamelCase_( unittest.TestCase ): @cached_property def snake_case__ ( self ): return ( AutoImageProcessor.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = SwinvaForImageClassification.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ).to( lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
708
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
709
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
0
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml __SCREAMING_SNAKE_CASE : Optional[int] = NewType('''DataClass''', Any) __SCREAMING_SNAKE_CASE : Dict = NewType('''DataClassType''', Any) def lowerCAmelCase_( lowercase_ : int ) -> int: if isinstance(lowercase_ , lowercase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase_( lowercase_ : list ) -> Callable[[str], Any]: _lowerCamelCase = {str(lowercase_ ): choice for choice in choices} return lambda lowercase_ : str_to_choice.get(lowercase_ , lowercase_ ) def lowerCAmelCase_( *, lowercase_ : Union[str, List[str]] = None , lowercase_ : str = None , lowercase_ : Any = dataclasses.MISSING , lowercase_ : Callable[[], Any] = dataclasses.MISSING , lowercase_ : dict = None , **lowercase_ : int , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls _lowerCamelCase = {} if aliases is not None: _lowerCamelCase = aliases if help is not None: _lowerCamelCase = help return dataclasses.field(metadata=lowercase_ , default=lowercase_ , default_factory=lowercase_ , **lowercase_ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Iterable[DataClassType] def __init__( self , lowerCamelCase__ , **lowerCamelCase__ ): # To make the default appear when using --help if "formatter_class" not in kwargs: _lowerCamelCase = ArgumentDefaultsHelpFormatter super().__init__(**lowerCamelCase__ ) if dataclasses.is_dataclass(lowerCamelCase__ ): _lowerCamelCase = [dataclass_types] _lowerCamelCase = list(lowerCamelCase__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(lowerCamelCase__ ) @staticmethod def snake_case__ ( lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = F"""--{field.name}""" _lowerCamelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , lowerCamelCase__ ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) _lowerCamelCase = kwargs.pop('''aliases''' , [] ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = [aliases] _lowerCamelCase = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(lowerCamelCase__ , '''UnionType''' ) and isinstance(lowerCamelCase__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(lowerCamelCase__ ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' F""" Problem encountered in field '{field.name}'.""" ) if type(lowerCamelCase__ ) not in field.type.__args__: # filter `str` in Union _lowerCamelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] _lowerCamelCase = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) _lowerCamelCase = ( field.type.__args__[0] if isinstance(lowerCamelCase__ , field.type.__args__[1] ) else field.type.__args__[1] ) _lowerCamelCase = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) _lowerCamelCase = {} if origin_type is Literal or (isinstance(field.type , lowerCamelCase__ ) and issubclass(field.type , lowerCamelCase__ )): if origin_type is Literal: _lowerCamelCase = field.type.__args__ else: _lowerCamelCase = [x.value for x in field.type] _lowerCamelCase = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: _lowerCamelCase = field.default else: _lowerCamelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument _lowerCamelCase = copy(lowerCamelCase__ ) # Hack because type=bool in argparse does not behave as we want. _lowerCamelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. _lowerCamelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way _lowerCamelCase = default # This tells argparse we accept 0 or 1 value after --field_name _lowerCamelCase = '''?''' # This is the value that will get picked if we do --field_name (without value) _lowerCamelCase = True elif isclass(lowerCamelCase__ ) and issubclass(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = field.type.__args__[0] _lowerCamelCase = '''+''' if field.default_factory is not dataclasses.MISSING: _lowerCamelCase = field.default_factory() elif field.default is dataclasses.MISSING: _lowerCamelCase = True else: _lowerCamelCase = field.type if field.default is not dataclasses.MISSING: _lowerCamelCase = field.default elif field.default_factory is not dataclasses.MISSING: _lowerCamelCase = field.default_factory() else: _lowerCamelCase = True parser.add_argument(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): _lowerCamelCase = False parser.add_argument(F"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): if hasattr(lowerCamelCase__ , '''_argument_group_name''' ): _lowerCamelCase = self.add_argument_group(dtype._argument_group_name ) else: _lowerCamelCase = self try: _lowerCamelCase = get_type_hints(lowerCamelCase__ ) except NameError: raise RuntimeError( F"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(lowerCamelCase__ ): _lowerCamelCase = '''.'''.join(map(lowerCamelCase__ , sys.version_info[:3] ) ) raise RuntimeError( F"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(lowerCamelCase__ ): if not field.init: continue _lowerCamelCase = type_hints[field.name] self._parse_dataclass_field(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=None , ): if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): _lowerCamelCase = [] if args_filename: args_files.append(Path(lowerCamelCase__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values _lowerCamelCase = ArgumentParser() args_file_parser.add_argument(lowerCamelCase__ , type=lowerCamelCase__ , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) _lowerCamelCase , _lowerCamelCase = args_file_parser.parse_known_args(args=lowerCamelCase__ ) _lowerCamelCase = vars(lowerCamelCase__ ).get(args_file_flag.lstrip('''-''' ) , lowerCamelCase__ ) if cmd_args_file_paths: args_files.extend([Path(lowerCamelCase__ ) for p in cmd_args_file_paths] ) _lowerCamelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last _lowerCamelCase = file_args + args if args is not None else file_args + sys.argv[1:] _lowerCamelCase , _lowerCamelCase = self.parse_known_args(args=lowerCamelCase__ ) _lowerCamelCase = [] for dtype in self.dataclass_types: _lowerCamelCase = {f.name for f in dataclasses.fields(lowerCamelCase__ ) if f.init} _lowerCamelCase = {k: v for k, v in vars(lowerCamelCase__ ).items() if k in keys} for k in keys: delattr(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = dtype(**lowerCamelCase__ ) outputs.append(lowerCamelCase__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(lowerCamelCase__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): _lowerCamelCase = set(args.keys() ) _lowerCamelCase = [] for dtype in self.dataclass_types: _lowerCamelCase = {f.name for f in dataclasses.fields(lowerCamelCase__ ) if f.init} _lowerCamelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) _lowerCamelCase = dtype(**lowerCamelCase__ ) outputs.append(lowerCamelCase__ ) if not allow_extra_keys and unused_keys: raise ValueError(F"""Some keys are not used by the HfArgumentParser: {sorted(lowerCamelCase__ )}""" ) return tuple(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): with open(Path(lowerCamelCase__ ) , encoding='''utf-8''' ) as open_json_file: _lowerCamelCase = json.loads(open_json_file.read() ) _lowerCamelCase = self.parse_dict(lowerCamelCase__ , allow_extra_keys=lowerCamelCase__ ) return tuple(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): _lowerCamelCase = self.parse_dict(yaml.safe_load(Path(lowerCamelCase__ ).read_text() ) , allow_extra_keys=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
710
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
0
"""simple docstring""" from scipy.stats import pearsonr import datasets __SCREAMING_SNAKE_CASE : Dict = ''' Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. ''' __SCREAMING_SNAKE_CASE : int = ''' Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results[\'pearsonr\'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) [\'p-value\', \'pearsonr\'] >>> print(round(results[\'pearsonr\'], 2)) -0.74 >>> print(round(results[\'p-value\'], 2)) 0.15 ''' __SCREAMING_SNAKE_CASE : int = ''' @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): if return_pvalue: _lowerCamelCase = pearsonr(lowerCamelCase__ , lowerCamelCase__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(lowerCamelCase__ , lowerCamelCase__ )[0] )}
711
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
0
"""simple docstring""" from __future__ import annotations from random import choice def lowerCAmelCase_( lowercase_ : Tuple ) -> Optional[int]: return choice(lowercase_ ) def lowerCAmelCase_( lowercase_ : list[int] , lowercase_ : int ) -> int: _lowerCamelCase = random_pivot(lowercase_ ) # partition based on pivot # linear time _lowerCamelCase = [e for e in lst if e < pivot] _lowerCamelCase = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(lowercase_ ) == k - 1: return pivot # pivot is in elements bigger than k elif len(lowercase_ ) < k - 1: return kth_number(lowercase_ , k - len(lowercase_ ) - 1 ) # pivot is in elements smaller than k else: return kth_number(lowercase_ , lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
712
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
713
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
0
from __future__ import annotations def lowerCAmelCase_( lowercase_ : int ) -> list[int]: _lowerCamelCase = 2 _lowerCamelCase = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(lowercase_ ) if n > 1: factors.append(lowercase_ ) return factors if __name__ == "__main__": import doctest doctest.testmod()
714
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Any = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Optional[Any] = 'megatron-bert' def __init__( self , lowerCamelCase__=2_9_0_5_6 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=4_0_9_6 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=0 , lowerCamelCase__="absolute" , lowerCamelCase__=True , **lowerCamelCase__ , ): super().__init__(pad_token_id=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = position_embedding_type _lowerCamelCase = use_cache
715
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
0
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 10_00 ) -> int: _lowerCamelCase , _lowerCamelCase = 1, 1 _lowerCamelCase = 2 while True: _lowerCamelCase = 0 _lowerCamelCase = fa + fa _lowerCamelCase , _lowerCamelCase = fa, f index += 1 for _ in str(lowercase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
716
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
623
0
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = BarthezTokenizer lowercase__ : List[str] = BarthezTokenizerFast lowercase__ : int = True lowercase__ : List[str] = True def snake_case__ ( self ): super().setUp() _lowerCamelCase = BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer def snake_case__ ( self ): _lowerCamelCase = '''<pad>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCamelCase__ ) , 1_0_1_1_2_2 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2 ) @require_torch def snake_case__ ( self ): _lowerCamelCase = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] _lowerCamelCase = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] _lowerCamelCase = self.tokenizer( lowerCamelCase__ , max_length=len(lowerCamelCase__ ) , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , return_tensors='''pt''' ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) _lowerCamelCase = batch.input_ids.tolist()[0] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # fmt: off _lowerCamelCase = {'''input_ids''': [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. _lowerCamelCase = [ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=lowerCamelCase__ , )
717
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
0
"""simple docstring""" import requests from bsa import BeautifulSoup def lowerCAmelCase_( lowercase_ : str , lowercase_ : dict ) -> str: _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , params=lowercase_ ).content , '''html.parser''' ) _lowerCamelCase = soup.find('''div''' , attrs={'''class''': '''gs_ri'''} ) _lowerCamelCase = div.find('''div''' , attrs={'''class''': '''gs_fl'''} ).find_all('''a''' ) return anchors[2].get_text() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = { '''title''': ( '''Precisely geometry controlled microsupercapacitors for ultrahigh areal ''' '''capacitance, volumetric capacitance, and energy density''' ), '''journal''': '''Chem. Mater.''', '''volume''': 3_0, '''pages''': '''3979-3990''', '''year''': 2_0_1_8, '''hl''': '''en''', } print(get_citation('''https://scholar.google.com/scholar_lookup''', params=params))
718
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
0
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Any , lowercase_ : List[str] ) -> Optional[int]: # Initialise PyTorch model _lowerCamelCase = MobileBertConfig.from_json_file(lowercase_ ) print(F"""Building PyTorch model from configuration: {config}""" ) _lowerCamelCase = MobileBertForPreTraining(lowercase_ ) # Load weights from tf checkpoint _lowerCamelCase = load_tf_weights_in_mobilebert(lowercase_ , lowercase_ , lowercase_ ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--mobilebert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained MobileBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
719
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
0
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''vocab_file''': '''spiece.model'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', } } __SCREAMING_SNAKE_CASE : Optional[Any] = { '''albert-base-v1''': 5_1_2, '''albert-large-v1''': 5_1_2, '''albert-xlarge-v1''': 5_1_2, '''albert-xxlarge-v1''': 5_1_2, '''albert-base-v2''': 5_1_2, '''albert-large-v2''': 5_1_2, '''albert-xlarge-v2''': 5_1_2, '''albert-xxlarge-v2''': 5_1_2, } __SCREAMING_SNAKE_CASE : Optional[Any] = '''▁''' class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = VOCAB_FILES_NAMES lowercase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowerCamelCase__ , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__="[CLS]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="<unk>" , lowerCamelCase__="[SEP]" , lowerCamelCase__="<pad>" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__ = None , **lowerCamelCase__ , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _lowerCamelCase = ( AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ , normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token ) _lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCamelCase__ , remove_space=lowerCamelCase__ , keep_accents=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) _lowerCamelCase = do_lower_case _lowerCamelCase = remove_space _lowerCamelCase = keep_accents _lowerCamelCase = vocab_file _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase__ ) @property def snake_case__ ( self ): return len(self.sp_model ) def snake_case__ ( self ): _lowerCamelCase = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): _lowerCamelCase = self.__dict__.copy() _lowerCamelCase = None return state def __setstate__( self , lowerCamelCase__ ): _lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _lowerCamelCase = {} _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def snake_case__ ( self , lowerCamelCase__ ): if self.remove_space: _lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: _lowerCamelCase = inputs _lowerCamelCase = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: _lowerCamelCase = unicodedata.normalize('''NFKD''' , lowerCamelCase__ ) _lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(lowerCamelCase__ )] ) if self.do_lower_case: _lowerCamelCase = outputs.lower() return outputs def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.preprocess_text(lowerCamelCase__ ) _lowerCamelCase = self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) _lowerCamelCase = [] for piece in pieces: if len(lowerCamelCase__ ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): _lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(lowerCamelCase__ , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _lowerCamelCase = cur_pieces[1:] else: _lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(lowerCamelCase__ ) else: new_pieces.append(lowerCamelCase__ ) return new_pieces def snake_case__ ( self , lowerCamelCase__ ): return self.sp_model.PieceToId(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): return self.sp_model.IdToPiece(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] _lowerCamelCase = '''''' _lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCamelCase__ ) + token _lowerCamelCase = True _lowerCamelCase = [] else: current_sub_tokens.append(lowerCamelCase__ ) _lowerCamelCase = False out_string += self.sp_model.decode(lowerCamelCase__ ) return out_string.strip() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is not None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , '''wb''' ) as fi: _lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,)
720
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
0
"""simple docstring""" from __future__ import annotations from random import random class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ = None ): _lowerCamelCase = value _lowerCamelCase = random() _lowerCamelCase = None _lowerCamelCase = None def __repr__( self ): from pprint import pformat if self.left is None and self.right is None: return F"""'{self.value}: {self.prior:.5}'""" else: return pformat( {F"""{self.value}: {self.prior:.5}""": (self.left, self.right)} , indent=1 ) def __str__( self ): _lowerCamelCase = str(self.value ) + ''' ''' _lowerCamelCase = str(self.left or '''''' ) _lowerCamelCase = str(self.right or '''''' ) return value + left + right def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : int ) -> tuple[Node | None, Node | None]: if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: _lowerCamelCase , _lowerCamelCase = split(root.left , lowercase_ ) return left, root else: _lowerCamelCase , _lowerCamelCase = split(root.right , lowercase_ ) return root, right def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : Node | None ) -> Node | None: if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: _lowerCamelCase = merge(left.right , lowercase_ ) return left else: _lowerCamelCase = merge(lowercase_ , right.left ) return right def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : int ) -> Node | None: _lowerCamelCase = Node(lowercase_ ) _lowerCamelCase , _lowerCamelCase = split(lowercase_ , lowercase_ ) return merge(merge(lowercase_ , lowercase_ ) , lowercase_ ) def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : int ) -> Node | None: _lowerCamelCase , _lowerCamelCase = split(lowercase_ , value - 1 ) _lowerCamelCase , _lowerCamelCase = split(lowercase_ , lowercase_ ) return merge(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : Node | None ) -> None: if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : str ) -> Node | None: for arg in args.split(): if arg[0] == "+": _lowerCamelCase = insert(lowercase_ , int(arg[1:] ) ) elif arg[0] == "-": _lowerCamelCase = erase(lowercase_ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def lowerCAmelCase_( ) -> None: _lowerCamelCase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) _lowerCamelCase = input() while args != "q": _lowerCamelCase = interact_treap(lowercase_ , lowercase_ ) print(lowercase_ ) _lowerCamelCase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
721
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
0
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
1
'''simple docstring''' def _A ( A__ = 1000 ): """simple docstring""" __lowercase = 2**power __lowercase = str(A__ ) __lowercase = list(A__ ) __lowercase = 0 for i in list_num: sum_of_num += int(A__ ) return sum_of_num if __name__ == "__main__": lowerCAmelCase__ = int(input('''Enter the power of 2: ''').strip()) print('''2 ^ ''', power, ''' = ''', 2**power) lowerCAmelCase__ = solution(power) print('''Sum of the digits is: ''', result)
624
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
1
'''simple docstring''' lowerCAmelCase__ = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' lowerCAmelCase__ = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] lowerCAmelCase__ = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
624
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
1
'''simple docstring''' def _A ( A__ ): """simple docstring""" __lowercase = len(A__ ) for i in range(length - 1 ): __lowercase = i for k in range(i + 1 , A__ ): if collection[k] < collection[least]: __lowercase = k if least != i: __lowercase , __lowercase = (collection[i], collection[least]) return collection if __name__ == "__main__": lowerCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(''',''')] print(selection_sort(unsorted))
624
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
1
'''simple docstring''' lowerCAmelCase__ = frozenset( [ '''prompt''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', '''cross_attention_kwargs''', ] ) lowerCAmelCase__ = frozenset(['''prompt''', '''negative_prompt''']) lowerCAmelCase__ = frozenset([]) lowerCAmelCase__ = frozenset(['''image''']) lowerCAmelCase__ = frozenset( [ '''image''', '''height''', '''width''', '''guidance_scale''', ] ) lowerCAmelCase__ = frozenset(['''image''']) lowerCAmelCase__ = frozenset( [ '''prompt''', '''image''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', ] ) lowerCAmelCase__ = frozenset(['''prompt''', '''image''', '''negative_prompt''']) lowerCAmelCase__ = frozenset( [ # Text guided image variation with an image mask '''prompt''', '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', ] ) lowerCAmelCase__ = frozenset(['''prompt''', '''image''', '''mask_image''', '''negative_prompt''']) lowerCAmelCase__ = frozenset( [ # image variation with an image mask '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', ] ) lowerCAmelCase__ = frozenset(['''image''', '''mask_image''']) lowerCAmelCase__ = frozenset( [ '''example_image''', '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', ] ) lowerCAmelCase__ = frozenset(['''example_image''', '''image''', '''mask_image''']) lowerCAmelCase__ = frozenset(['''class_labels''']) lowerCAmelCase__ = frozenset(['''class_labels''']) lowerCAmelCase__ = frozenset(['''batch_size''']) lowerCAmelCase__ = frozenset([]) lowerCAmelCase__ = frozenset(['''batch_size''']) lowerCAmelCase__ = frozenset([]) lowerCAmelCase__ = frozenset( [ '''prompt''', '''audio_length_in_s''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', '''cross_attention_kwargs''', ] ) lowerCAmelCase__ = frozenset(['''prompt''', '''negative_prompt''']) lowerCAmelCase__ = frozenset(['''input_tokens''']) lowerCAmelCase__ = frozenset(['''input_tokens'''])
624
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
1
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
624
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : Optional[int] ,lowercase__ : List[str] ,lowercase__ : List[Any]=7 ,lowercase__ : int=3 ,lowercase__ : Dict=1_8 ,lowercase__ : Optional[int]=3_0 ,lowercase__ : str=4_0_0 ,lowercase__ : Optional[Any]=True ,lowercase__ : Any=None ,lowercase__ : Any=True ,lowercase__ : Tuple=None ,): __lowercase = size if size is not None else {'''shortest_edge''': 2_0} __lowercase = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = do_center_crop __lowercase = crop_size def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = MobileNetVaImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = MobileNetVaImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : str ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase__ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''size''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_center_crop''' ) ) self.assertTrue(hasattr(lowercase__ ,'''crop_size''' ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size ,{'''height''': 1_8, '''width''': 1_8} ) __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ,crop_size=8_4 ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size ,{'''height''': 8_4, '''width''': 8_4} ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,numpify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Tuple ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,torchify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,)
624
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: lowerCAmelCase__ = None lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } lowerCAmelCase__ = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } lowerCAmelCase__ = '''▁''' class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : int = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Tuple = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Tuple = AlbertTokenizer def __init__( self : Any ,lowercase__ : Tuple=None ,lowercase__ : str=None ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=True ,lowercase__ : Optional[Any]=False ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : Optional[Any]="[SEP]" ,lowercase__ : Any="<unk>" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : Optional[int]="<pad>" ,lowercase__ : Tuple="[CLS]" ,lowercase__ : int="[MASK]" ,**lowercase__ : List[Any] ,): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __lowercase = ( AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ,normalized=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else mask_token ) super().__init__( lowercase__ ,tokenizer_file=lowercase__ ,do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = False if not self.vocab_file else True def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(lowercase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
1
'''simple docstring''' from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : str = 'EncodecFeatureExtractor' SCREAMING_SNAKE_CASE : List[Any] = ('T5Tokenizer', 'T5TokenizerFast') def __init__( self : str ,lowercase__ : int ,lowercase__ : Optional[Any] ): super().__init__(lowercase__ ,lowercase__ ) __lowercase = self.feature_extractor __lowercase = False def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any=None ,lowercase__ : Any=None ,lowercase__ : Tuple=True ): return self.tokenizer.get_decoder_prompt_ids(task=lowercase__ ,language=lowercase__ ,no_timestamps=lowercase__ ) def __call__( self : Any ,*lowercase__ : Dict ,**lowercase__ : int ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*lowercase__ ,**lowercase__ ) __lowercase = kwargs.pop('''audio''' ,lowercase__ ) __lowercase = kwargs.pop('''sampling_rate''' ,lowercase__ ) __lowercase = kwargs.pop('''text''' ,lowercase__ ) if len(lowercase__ ) > 0: __lowercase = args[0] __lowercase = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if text is not None: __lowercase = self.tokenizer(lowercase__ ,**lowercase__ ) if audio is not None: __lowercase = self.feature_extractor(lowercase__ ,*lowercase__ ,sampling_rate=lowercase__ ,**lowercase__ ) if audio is None: return inputs elif text is None: return audio_inputs else: __lowercase = audio_inputs['''input_values'''] if "padding_mask" in audio_inputs: __lowercase = audio_inputs['''padding_mask'''] return inputs def SCREAMING_SNAKE_CASE ( self : Any ,*lowercase__ : Optional[Any] ,**lowercase__ : Optional[int] ): __lowercase = kwargs.pop('''audio''' ,lowercase__ ) __lowercase = kwargs.pop('''padding_mask''' ,lowercase__ ) if len(lowercase__ ) > 0: __lowercase = args[0] __lowercase = args[1:] if audio_values is not None: return self._decode_audio(lowercase__ ,padding_mask=lowercase__ ) else: return self.tokenizer.batch_decode(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,*lowercase__ : int ,**lowercase__ : List[Any] ): return self.tokenizer.decode(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : List[str] ,lowercase__ : Optional = None ): __lowercase = to_numpy(lowercase__ ) __lowercase , __lowercase , __lowercase = audio_values.shape if padding_mask is None: return list(lowercase__ ) __lowercase = to_numpy(lowercase__ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) __lowercase = seq_len - padding_mask.shape[-1] __lowercase = 1 - self.feature_extractor.padding_value __lowercase = np.pad(lowercase__ ,((0, 0), (0, difference)) ,'''constant''' ,constant_values=lowercase__ ) __lowercase = audio_values.tolist() for i in range(lowercase__ ): __lowercase = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] __lowercase = sliced_audio.reshape(lowercase__ ,-1 ) return audio_values
624
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
1
'''simple docstring''' import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline lowerCAmelCase__ = { '''n_samples''': 64, '''horizon''': 32, '''num_inference_steps''': 20, '''n_guide_steps''': 2, # can set to 0 for faster sampling, does not use value network '''scale_grad_by_std''': True, '''scale''': 0.1, '''eta''': 0.0, '''t_grad_cutoff''': 2, '''device''': '''cpu''', } if __name__ == "__main__": lowerCAmelCase__ = '''hopper-medium-v2''' lowerCAmelCase__ = gym.make(env_name) lowerCAmelCase__ = ValueGuidedRLPipeline.from_pretrained( '''bglick13/hopper-medium-v2-value-function-hor32''', env=env, ) env.seed(0) lowerCAmelCase__ = env.reset() lowerCAmelCase__ = 0 lowerCAmelCase__ = 0 lowerCAmelCase__ = 1000 lowerCAmelCase__ = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy lowerCAmelCase__ = pipeline(obs, planning_horizon=32) # execute action in environment lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = env.step(denorm_actions) lowerCAmelCase__ = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( f'Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:' f' {total_score}' ) # save observations for rendering rollout.append(next_observation.copy()) lowerCAmelCase__ = next_observation except KeyboardInterrupt: pass print(f'Total reward: {total_reward}')
624
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
1
'''simple docstring''' import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ).convert('''RGB''' ) __lowercase = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3) , (0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1) ), ] ) __lowercase = transform(A__ ).unsqueeze(0 ).to(A__ ) return image def _A ( A__ ): """simple docstring""" if "visual_encoder" in key: __lowercase = re.sub('''visual_encoder*''' , '''vision_model.encoder''' , A__ ) if "blocks" in key: __lowercase = re.sub(R'''blocks''' , '''layers''' , A__ ) if "attn" in key: __lowercase = re.sub(R'''attn''' , '''self_attn''' , A__ ) if "norm1" in key: __lowercase = re.sub(R'''norm1''' , '''layer_norm1''' , A__ ) if "norm2" in key: __lowercase = re.sub(R'''norm2''' , '''layer_norm2''' , A__ ) if "encoder.norm" in key: __lowercase = re.sub(R'''encoder.norm''' , '''post_layernorm''' , A__ ) if "encoder.patch_embed.proj" in key: __lowercase = re.sub(R'''encoder.patch_embed.proj''' , '''embeddings.patch_embedding''' , A__ ) if "encoder.pos_embed" in key: __lowercase = re.sub(R'''encoder.pos_embed''' , '''embeddings.position_embedding''' , A__ ) if "encoder.cls_token" in key: __lowercase = re.sub(R'''encoder.cls_token''' , '''embeddings.class_embedding''' , A__ ) if "self_attn" in key: __lowercase = re.sub(R'''self_attn.proj''' , '''self_attn.projection''' , A__ ) return key @torch.no_grad() def _A ( A__ , A__=None ): """simple docstring""" if config_path is not None: __lowercase = BlipConfig.from_pretrained(A__ ) else: __lowercase = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) __lowercase = BlipForConditionalGeneration(A__ ).eval() __lowercase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth''' __lowercase = blip_decoder(pretrained=A__ , image_size=384 , vit='''base''' ) __lowercase = pt_model.eval() __lowercase = pt_model.state_dict() for key in modified_state_dict.copy(): __lowercase = modified_state_dict.pop(A__ ) __lowercase = rename_key(A__ ) __lowercase = value hf_model.load_state_dict(A__ ) __lowercase = 384 __lowercase = load_demo_image(image_size=A__ , device='''cpu''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) __lowercase = tokenizer(['''a picture of'''] ).input_ids __lowercase = hf_model.generate(A__ , A__ ) assert out[0].tolist() == [30522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] __lowercase = hf_model.generate(A__ ) assert out[0].tolist() == [30522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(A__ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' __lowercase = ( '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth''' ) __lowercase = blip_vqa(pretrained=A__ , image_size=A__ , vit='''base''' ) vqa_model.eval() __lowercase = vqa_model.state_dict() for key in modified_state_dict.copy(): __lowercase = modified_state_dict.pop(A__ ) __lowercase = rename_key(A__ ) __lowercase = value __lowercase = BlipForQuestionAnswering(A__ ) hf_vqa_model.load_state_dict(A__ ) __lowercase = ['''How many dogs are in this image?'''] __lowercase = tokenizer(A__ , return_tensors='''pt''' ).input_ids __lowercase = hf_vqa_model.generate(A__ , A__ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '''_vqa''' ) __lowercase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth''' __lowercase = blip_itm(pretrained=A__ , image_size=A__ , vit='''base''' ) itm_model.eval() __lowercase = itm_model.state_dict() for key in modified_state_dict.copy(): __lowercase = modified_state_dict.pop(A__ ) __lowercase = rename_key(A__ ) __lowercase = value __lowercase = BlipForImageTextRetrieval(A__ ) __lowercase = ['''A picture of a woman with a dog sitting in a beach'''] __lowercase = tokenizer( A__ , return_tensors='''pt''' , padding='''max_length''' , truncation=A__ , max_length=35 , ).input_ids hf_itm_model.load_state_dict(A__ ) hf_itm_model.eval() __lowercase = hf_itm_model(A__ , A__ , use_itm_head=A__ ) __lowercase = hf_itm_model(A__ , A__ , use_itm_head=A__ ) assert out[0].item() == 0.2_1_1_0_6_8_7_4_9_4_2_7_7_9_5_4 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_5_6_9_8_8_4_5_3_8_6_5_0_5_1_2_7 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '''_itm''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCAmelCase__ = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
624
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ ): """simple docstring""" __lowercase = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: __lowercase = 1024 __lowercase = 4096 __lowercase = 24 __lowercase = 16 __lowercase = [5, 11, 17, 23] __lowercase = [256, 512, 1024, 1024] __lowercase = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: __lowercase = 768 __lowercase = [1, 1, 1, 0.5] __lowercase = [256, 512, 768, 768] __lowercase = 150 __lowercase = 16 __lowercase = (1, 384, 384) __lowercase = False __lowercase = '''project''' if "ade" in checkpoint_url: __lowercase = True __lowercase = 768 __lowercase = [1, 1, 1, 0.5] __lowercase = 150 __lowercase = 16 __lowercase = '''huggingface/label-files''' __lowercase = '''ade20k-id2label.json''' __lowercase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type='''dataset''' ) ) , '''r''' ) ) __lowercase = {int(A__ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} __lowercase = [1, 150, 480, 480] return config, expected_shape def _A ( A__ ): """simple docstring""" __lowercase = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(A__ , A__ ) def _A ( A__ ): """simple docstring""" if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowercase = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: __lowercase = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: __lowercase = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: __lowercase = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: __lowercase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: __lowercase = name.replace('''proj''' , '''projection''' ) if "blocks" in name: __lowercase = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: __lowercase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: __lowercase = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: __lowercase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: __lowercase = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: __lowercase = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: __lowercase = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: __lowercase = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: __lowercase = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: __lowercase = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: __lowercase = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: __lowercase = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowercase = name.replace(F"refinenet{layer_idx}" , F"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: __lowercase = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: __lowercase = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: __lowercase = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: __lowercase = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: __lowercase = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowercase = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: __lowercase = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: __lowercase = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: __lowercase = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: __lowercase = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: __lowercase = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: __lowercase = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: __lowercase = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: __lowercase = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: __lowercase = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: __lowercase = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: __lowercase = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: __lowercase = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: __lowercase = name.replace('''..''' , '''.''' ) if "stem.conv" in name: __lowercase = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: __lowercase = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: __lowercase = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: __lowercase = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: __lowercase = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: __lowercase = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: __lowercase = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def _A ( A__ , A__ ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowercase = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.weight" ) __lowercase = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase = in_proj_weight[: config.hidden_size, :] __lowercase = in_proj_bias[: config.hidden_size] __lowercase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowercase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowercase = in_proj_weight[ -config.hidden_size :, : ] __lowercase = in_proj_bias[-config.hidden_size :] def _A ( ): """simple docstring""" __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase , __lowercase = get_dpt_config(A__ ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") __lowercase = torch.load(A__ , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowercase = state_dict.pop(A__ ) __lowercase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowercase = DPTForSemanticSegmentation(A__ ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowercase = 480 if '''ade''' in checkpoint_url else 384 __lowercase = DPTImageProcessor(size=A__ ) __lowercase = prepare_img() __lowercase = image_processor(A__ , return_tensors='''pt''' ) # forward pass __lowercase = model(**A__ ).logits if '''ade''' in checkpoint_url else model(**A__ ).predicted_depth if show_prediction: __lowercase = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=A__ , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(A__ ).mkdir(exist_ok=A__ ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(A__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(A__ ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=False, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) parser.add_argument( '''--show_prediction''', action='''store_true''', ) lowerCAmelCase__ = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
624
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
1
'''simple docstring''' class lowercase_ : """simple docstring""" def __init__( self : Any ): __lowercase = 0 __lowercase = 0 __lowercase = {} def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): if vertex not in self.adjacency: __lowercase = {} self.num_vertices += 1 def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : Union[str, Any] ,lowercase__ : Any ): self.add_vertex(lowercase__ ) self.add_vertex(lowercase__ ) if head == tail: return __lowercase = weight __lowercase = weight def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.get_edges() for edge in edges: __lowercase , __lowercase , __lowercase = edge edges.remove((tail, head, weight) ) for i in range(len(lowercase__ ) ): __lowercase = list(edges[i] ) edges.sort(key=lambda lowercase__ : e[2] ) for i in range(len(lowercase__ ) - 1 ): if edges[i][2] >= edges[i + 1][2]: __lowercase = edges[i][2] + 1 for edge in edges: __lowercase , __lowercase , __lowercase = edge __lowercase = weight __lowercase = weight def __str__( self : Tuple ): __lowercase = '''''' for tail in self.adjacency: for head in self.adjacency[tail]: __lowercase = self.adjacency[head][tail] string += F"{head} -> {tail} == {weight}\n" return string.rstrip('''\n''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return self.adjacency.keys() @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : Any=None ,lowercase__ : List[str]=None ): __lowercase = Graph() if vertices is None: __lowercase = [] if edges is None: __lowercase = [] for vertex in vertices: g.add_vertex(lowercase__ ) for edge in edges: g.add_edge(*lowercase__ ) return g class lowercase_ : """simple docstring""" def __init__( self : Tuple ): __lowercase = {} __lowercase = {} def __len__( self : Dict ): return len(self.parent ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : str ): if item in self.parent: return self.find(lowercase__ ) __lowercase = item __lowercase = 0 return item def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : int ): if item not in self.parent: return self.make_set(lowercase__ ) if item != self.parent[item]: __lowercase = self.find(self.parent[item] ) return self.parent[item] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[str] ,lowercase__ : Optional[int] ): __lowercase = self.find(lowercase__ ) __lowercase = self.find(lowercase__ ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: __lowercase = roota return roota if self.rank[roota] < self.rank[roota]: __lowercase = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 __lowercase = roota return roota return None @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : Optional[Any] ): __lowercase = graph.num_vertices __lowercase = Graph.UnionFind() __lowercase = [] while num_components > 1: __lowercase = {} for vertex in graph.get_vertices(): __lowercase = -1 __lowercase = graph.get_edges() for edge in edges: __lowercase , __lowercase , __lowercase = edge edges.remove((tail, head, weight) ) for edge in edges: __lowercase , __lowercase , __lowercase = edge __lowercase = union_find.find(lowercase__ ) __lowercase = union_find.find(lowercase__ ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: __lowercase = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: __lowercase = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: __lowercase , __lowercase , __lowercase = cheap_edge[vertex] if union_find.find(lowercase__ ) != union_find.find(lowercase__ ): union_find.union(lowercase__ ,lowercase__ ) mst_edges.append(cheap_edge[vertex] ) __lowercase = num_components - 1 __lowercase = Graph.build(edges=lowercase__ ) return mst
624
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
1
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" return int(input_a == input_a == 0 ) def _A ( ): """simple docstring""" print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(F"| 0 | 0 | {nor_gate(0 , 0 )} |" ) print(F"| 0 | 1 | {nor_gate(0 , 1 )} |" ) print(F"| 1 | 0 | {nor_gate(1 , 0 )} |" ) print(F"| 1 | 1 | {nor_gate(1 , 1 )} |" ) if __name__ == "__main__": import doctest doctest.testmod() main()
624
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = KandinskyVaaImgaImgPipeline SCREAMING_SNAKE_CASE : int = ['image_embeds', 'negative_image_embeds', 'image'] SCREAMING_SNAKE_CASE : List[str] = [ 'image_embeds', 'negative_image_embeds', 'image', ] SCREAMING_SNAKE_CASE : List[str] = [ 'generator', 'height', 'width', 'strength', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] SCREAMING_SNAKE_CASE : Tuple = False @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return 3_2 @property def SCREAMING_SNAKE_CASE ( self : Dict ): return 3_2 @property def SCREAMING_SNAKE_CASE ( self : List[str] ): return self.time_input_dim @property def SCREAMING_SNAKE_CASE ( self : List[Any] ): return self.time_input_dim * 4 @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return 1_0_0 @property def SCREAMING_SNAKE_CASE ( self : List[str] ): torch.manual_seed(0 ) __lowercase = { '''in_channels''': 4, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } __lowercase = UNetaDConditionModel(**lowercase__ ) return model @property def SCREAMING_SNAKE_CASE ( self : Tuple ): return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def SCREAMING_SNAKE_CASE ( self : str ): torch.manual_seed(0 ) __lowercase = VQModel(**self.dummy_movq_kwargs ) return model def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.dummy_unet __lowercase = self.dummy_movq __lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_schedule''': '''linear''', '''beta_start''': 0.0_0_0_8_5, '''beta_end''': 0.0_1_2, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } __lowercase = DDIMScheduler(**lowercase__ ) __lowercase = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : Tuple=0 ): __lowercase = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(lowercase__ ) ).to(lowercase__ ) __lowercase = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( lowercase__ ) # create init_image __lowercase = floats_tensor((1, 3, 6_4, 6_4) ,rng=random.Random(lowercase__ ) ).to(lowercase__ ) __lowercase = image.cpu().permute(0 ,2 ,3 ,1 )[0] __lowercase = Image.fromarray(np.uinta(lowercase__ ) ).convert('''RGB''' ).resize((2_5_6, 2_5_6) ) if str(lowercase__ ).startswith('''mps''' ): __lowercase = torch.manual_seed(lowercase__ ) else: __lowercase = torch.Generator(device=lowercase__ ).manual_seed(lowercase__ ) __lowercase = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''generator''': generator, '''height''': 6_4, '''width''': 6_4, '''num_inference_steps''': 1_0, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = '''cpu''' __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**lowercase__ ) __lowercase = pipe.to(lowercase__ ) pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = pipe(**self.get_dummy_inputs(lowercase__ ) ) __lowercase = output.images __lowercase = pipe( **self.get_dummy_inputs(lowercase__ ) ,return_dict=lowercase__ ,)[0] __lowercase = image[0, -3:, -3:, -1] __lowercase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) __lowercase = np.array( [0.6_1_9_9_7_7_8, 0.6_3_9_8_4_4_0_6, 0.4_6_1_4_5_7_8_5, 0.6_2_9_4_4_9_8_4, 0.5_6_2_2_2_1_5, 0.4_7_3_0_6_1_3_2, 0.4_7_4_4_1_4_5_6, 0.4_6_0_7_6_0_6, 0.4_8_7_1_9_2_6_3] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_img2img_frog.npy''' ) __lowercase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) __lowercase = '''A red cartoon frog, 4k''' __lowercase = KandinskyVaaPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' ,torch_dtype=torch.floataa ) pipe_prior.to(lowercase__ ) __lowercase = KandinskyVaaImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-decoder''' ,torch_dtype=torch.floataa ) __lowercase = pipeline.to(lowercase__ ) pipeline.set_progress_bar_config(disable=lowercase__ ) __lowercase = torch.Generator(device='''cpu''' ).manual_seed(0 ) __lowercase , __lowercase = pipe_prior( lowercase__ ,generator=lowercase__ ,num_inference_steps=5 ,negative_prompt='''''' ,).to_tuple() __lowercase = pipeline( image=lowercase__ ,image_embeds=lowercase__ ,negative_image_embeds=lowercase__ ,generator=lowercase__ ,num_inference_steps=1_0_0 ,height=7_6_8 ,width=7_6_8 ,strength=0.2 ,output_type='''np''' ,) __lowercase = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(lowercase__ ,lowercase__ )
624
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
1
'''simple docstring''' from collections import defaultdict def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = True for v in tree[start]: if v not in visited: ret += dfs(A__ ) if ret % 2 == 0: cuts.append(A__ ) return ret def _A ( ): """simple docstring""" dfs(1 ) if __name__ == "__main__": lowerCAmelCase__ , lowerCAmelCase__ = 10, 9 lowerCAmelCase__ = defaultdict(list) lowerCAmelCase__ = {} lowerCAmelCase__ = [] lowerCAmelCase__ = 0 lowerCAmelCase__ = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
624
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
1
'''simple docstring''' from math import factorial def _A ( A__ , A__ ): """simple docstring""" if n < k or k < 0: raise ValueError('''Please enter positive integers for n and k where n >= k''' ) return factorial(A__ ) // (factorial(A__ ) * factorial(n - k )) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f'fifty-two card deck is: {combinations(52, 5)}\n', ) print( '''If a class of 40 students must be arranged into groups of''', f'4 for group projects, there are {combinations(40, 4)} ways', '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f'are {combinations(10, 3)} ways that first, second and', '''third place can be awarded.''', )
624
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
1
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _A ( A__=None ): """simple docstring""" if subparsers is not None: __lowercase = subparsers.add_parser('''test''' ) else: __lowercase = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=A__ , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=A__ ) return parser def _A ( A__ ): """simple docstring""" __lowercase = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: __lowercase = script_name else: __lowercase = F"--config_file={args.config_file} {script_name}" __lowercase = ['''accelerate-launch'''] + test_args.split() __lowercase = execute_subprocess_async(A__ , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def _A ( ): """simple docstring""" __lowercase = test_command_parser() __lowercase = parser.parse_args() test_command(A__ ) if __name__ == "__main__": main()
624
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
1
'''simple docstring''' from math import factorial class lowercase_ : """simple docstring""" def __init__( self : Dict ,lowercase__ : str ,lowercase__ : Tuple ): __lowercase = real if isinstance(lowercase__ ,lowercase__ ): __lowercase = [1] * rank else: __lowercase = rank def __repr__( self : Tuple ): return ( F"{self.real}+" F"{'+'.join(str(lowercase__ )+'E'+str(n+1 )for n,dual in enumerate(self.duals ) )}" ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real ,lowercase__ ) def __add__( self : Optional[Any] ,lowercase__ : str ): if not isinstance(lowercase__ ,lowercase__ ): return Dual(self.real + other ,self.duals ) __lowercase = self.duals.copy() __lowercase = other.duals.copy() if len(lowercase__ ) > len(lowercase__ ): o_dual.extend([1] * (len(lowercase__ ) - len(lowercase__ )) ) elif len(lowercase__ ) < len(lowercase__ ): s_dual.extend([1] * (len(lowercase__ ) - len(lowercase__ )) ) __lowercase = [] for i in range(len(lowercase__ ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real ,lowercase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = __add__ def __sub__( self : Union[str, Any] ,lowercase__ : int ): return self + other * -1 def __mul__( self : Dict ,lowercase__ : Optional[int] ): if not isinstance(lowercase__ ,lowercase__ ): __lowercase = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other ,lowercase__ ) __lowercase = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real ,lowercase__ ) SCREAMING_SNAKE_CASE : Dict = __mul__ def __truediv__( self : Tuple ,lowercase__ : Dict ): if not isinstance(lowercase__ ,lowercase__ ): __lowercase = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other ,lowercase__ ) raise ValueError def __floordiv__( self : List[str] ,lowercase__ : Any ): if not isinstance(lowercase__ ,lowercase__ ): __lowercase = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other ,lowercase__ ) raise ValueError def __pow__( self : Union[str, Any] ,lowercase__ : Dict ): if n < 0 or isinstance(lowercase__ ,lowercase__ ): raise ValueError('''power must be a positive integer''' ) if n == 0: return 1 if n == 1: return self __lowercase = self for _ in range(n - 1 ): x *= self return x def _A ( A__ , A__ , A__ ): """simple docstring""" if not callable(A__ ): raise ValueError('''differentiate() requires a function as input for func''' ) if not isinstance(A__ , (float, int) ): raise ValueError('''differentiate() requires a float as input for position''' ) if not isinstance(A__ , A__ ): raise ValueError('''differentiate() requires an int as input for order''' ) __lowercase = Dual(A__ , 1 ) __lowercase = func(A__ ) if order == 0: return result.real return result.duals[order - 1] * factorial(A__ ) if __name__ == "__main__": import doctest doctest.testmod() def _A ( A__ ): """simple docstring""" return y**2 * y**4 print(differentiate(f, 9, 2))
624
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
1
'''simple docstring''' from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent lowerCAmelCase__ = {'''UserAgent''': UserAgent().random} def _A ( A__ ): """simple docstring""" __lowercase = script.contents[0] __lowercase = json.loads(data[data.find('''{"config"''' ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : Optional[int] ): __lowercase = F"https://www.instagram.com/{username}/" __lowercase = self.get_json() def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = requests.get(self.url ,headers=lowercase__ ).text __lowercase = BeautifulSoup(lowercase__ ,'''html.parser''' ).find_all('''script''' ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self : Optional[int] ): return F"{self.__class__.__name__}('{self.username}')" def __str__( self : List[str] ): return F"{self.fullname} ({self.username}) is {self.biography}" @property def SCREAMING_SNAKE_CASE ( self : Tuple ): return self.user_data["username"] @property def SCREAMING_SNAKE_CASE ( self : int ): return self.user_data["full_name"] @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.user_data["biography"] @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.user_data["business_email"] @property def SCREAMING_SNAKE_CASE ( self : List[str] ): return self.user_data["external_url"] @property def SCREAMING_SNAKE_CASE ( self : List[str] ): return self.user_data["edge_followed_by"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Dict ): return self.user_data["edge_follow"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return self.user_data["edge_owner_to_timeline_media"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return self.user_data["profile_pic_url_hd"] @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return self.user_data["is_verified"] @property def SCREAMING_SNAKE_CASE ( self : Tuple ): return self.user_data["is_private"] def _A ( A__ = "github" ): """simple docstring""" import os if os.environ.get('''CI''' ): return # test failing on GitHub Actions __lowercase = InstagramUser(A__ ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , A__ ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 150 assert instagram_user.number_of_followers > 120000 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith('''https://instagram.''' ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = InstagramUser('''github''') print(instagram_user) print(f'{instagram_user.number_of_posts = }') print(f'{instagram_user.number_of_followers = }') print(f'{instagram_user.number_of_followings = }') print(f'{instagram_user.email = }') print(f'{instagram_user.website = }') print(f'{instagram_user.profile_picture_url = }') print(f'{instagram_user.is_verified = }') print(f'{instagram_user.is_private = }')
624
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
1
'''simple docstring''' from __future__ import annotations lowerCAmelCase__ = [] def _A ( A__ , A__ , A__ ): """simple docstring""" for i in range(len(A__ ) ): if board[row][i] == 1: return False for i in range(len(A__ ) ): if board[i][column] == 1: return False for i, j in zip(range(A__ , -1 , -1 ) , range(A__ , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(A__ , -1 , -1 ) , range(A__ , len(A__ ) ) ): if board[i][j] == 1: return False return True def _A ( A__ , A__ ): """simple docstring""" if row >= len(A__ ): solution.append(A__ ) printboard(A__ ) print() return True for i in range(len(A__ ) ): if is_safe(A__ , A__ , A__ ): __lowercase = 1 solve(A__ , row + 1 ) __lowercase = 0 return False def _A ( A__ ): """simple docstring""" for i in range(len(A__ ) ): for j in range(len(A__ ) ): if board[i][j] == 1: print('''Q''' , end=''' ''' ) else: print('''.''' , end=''' ''' ) print() # n=int(input("The no. of queens")) lowerCAmelCase__ = 8 lowerCAmelCase__ = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('''The total no. of solutions are :''', len(solution))
624
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: lowerCAmelCase__ = None lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/bigbird-roberta-base''': '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model''', '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model''' ), }, '''tokenizer_file''': { '''google/bigbird-roberta-base''': ( '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json''' ), '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase__ = { '''google/bigbird-roberta-base''': 4096, '''google/bigbird-roberta-large''': 4096, '''google/bigbird-base-trivia-itc''': 4096, } lowerCAmelCase__ = '''▁''' class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Dict = BigBirdTokenizer SCREAMING_SNAKE_CASE : Tuple = ['input_ids', 'attention_mask'] SCREAMING_SNAKE_CASE : List[int] = [] def __init__( self : List[str] ,lowercase__ : Optional[int]=None ,lowercase__ : Any=None ,lowercase__ : Optional[Any]="<unk>" ,lowercase__ : List[Any]="<s>" ,lowercase__ : Any="</s>" ,lowercase__ : str="<pad>" ,lowercase__ : Optional[int]="[SEP]" ,lowercase__ : Optional[Any]="[MASK]" ,lowercase__ : List[str]="[CLS]" ,**lowercase__ : List[Any] ,): __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else bos_token __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else eos_token __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else unk_token __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else pad_token __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else cls_token __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else mask_token super().__init__( lowercase__ ,tokenizer_file=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = vocab_file __lowercase = False if not self.vocab_file else True def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(lowercase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
1
'''simple docstring''' import math from collections.abc import Iterator from itertools import takewhile def _A ( A__ ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(A__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _A ( ): """simple docstring""" __lowercase = 2 while True: if is_prime(A__ ): yield num num += 1 def _A ( A__ = 2000000 ): """simple docstring""" return sum(takewhile(lambda A__ : x < n , prime_generator() ) ) if __name__ == "__main__": print(f'{solution() = }')
624
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
1
'''simple docstring''' import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = BioGptTokenizer SCREAMING_SNAKE_CASE : Optional[int] = False def SCREAMING_SNAKE_CASE ( self : str ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] __lowercase = dict(zip(lowercase__ ,range(len(lowercase__ ) ) ) ) __lowercase = ['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ) as fp: fp.write(json.dumps(lowercase__ ) ) with open(self.merges_file ,'''w''' ) as fp: fp.write('''\n'''.join(lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Optional[Any] ): __lowercase = '''lower newer''' __lowercase = '''lower newer''' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = BioGptTokenizer(self.vocab_file ,self.merges_file ) __lowercase = '''lower''' __lowercase = ['''low''', '''er</w>'''] __lowercase = tokenizer.tokenize(lowercase__ ) self.assertListEqual(lowercase__ ,lowercase__ ) __lowercase = tokens + ['''<unk>'''] __lowercase = [1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase__ ) ,lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = BioGptTokenizer.from_pretrained('''microsoft/biogpt''' ) __lowercase = tokenizer.encode('''sequence builders''' ,add_special_tokens=lowercase__ ) __lowercase = tokenizer.encode('''multi-sequence build''' ,add_special_tokens=lowercase__ ) __lowercase = tokenizer.build_inputs_with_special_tokens(lowercase__ ) __lowercase = tokenizer.build_inputs_with_special_tokens(lowercase__ ,lowercase__ ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
624
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
1
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
1
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) __lowercase = str(bin(A__ ) ) binary_number += "0" * shift_amount return binary_number def _A ( A__ , A__ ): """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) __lowercase = str(bin(A__ ) )[2:] if shift_amount >= len(A__ ): return "0b0" __lowercase = binary_number[: len(A__ ) - shift_amount] return "0b" + shifted_binary_number def _A ( A__ , A__ ): """simple docstring""" if number >= 0: # Get binary representation of positive number __lowercase = '''0''' + str(bin(A__ ) ).strip('''-''' )[2:] else: # Get binary (2's complement) representation of negative number __lowercase = len(bin(A__ )[3:] ) # Find 2's complement of number __lowercase = bin(abs(A__ ) - (1 << binary_number_length) )[3:] __lowercase = ( '''1''' + '''0''' * (binary_number_length - len(A__ )) + binary_number ) if shift_amount >= len(A__ ): return "0b" + binary_number[0] * len(A__ ) return ( "0b" + binary_number[0] * shift_amount + binary_number[: len(A__ ) - shift_amount] ) if __name__ == "__main__": import doctest doctest.testmod()
624
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
1
'''simple docstring''' import os from collections import deque import torch from torch.utils.data import Dataset class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Dict="" ,lowercase__ : Optional[int]="train" ): assert os.path.isdir(lowercase__ ) __lowercase = [] __lowercase = os.listdir(lowercase__ ) for story_filename in story_filenames_list: if "summary" in story_filename: continue __lowercase = os.path.join(lowercase__ ,lowercase__ ) if not os.path.isfile(lowercase__ ): continue self.documents.append(lowercase__ ) def __len__( self : Optional[int] ): return len(self.documents ) def __getitem__( self : Optional[Any] ,lowercase__ : Tuple ): __lowercase = self.documents[idx] __lowercase = document_path.split('''/''' )[-1] with open(lowercase__ ,encoding='''utf-8''' ) as source: __lowercase = source.read() __lowercase , __lowercase = process_story(lowercase__ ) return document_name, story_lines, summary_lines def _A ( A__ ): """simple docstring""" __lowercase = list(filter(lambda A__ : len(A__ ) != 0 , [line.strip() for line in raw_story.split('''\n''' )] ) ) # for some unknown reason some lines miss a period, add it __lowercase = [_add_missing_period(A__ ) for line in nonempty_lines] # gather article lines __lowercase = [] __lowercase = deque(A__ ) while True: try: __lowercase = lines.popleft() if element.startswith('''@highlight''' ): break story_lines.append(A__ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines __lowercase = list(filter(lambda A__ : not t.startswith('''@highlight''' ) , A__ ) ) return story_lines, summary_lines def _A ( A__ ): """simple docstring""" __lowercase = ['''.''', '''!''', '''?''', '''...''', '''\'''', '''`''', '''"''', '''\u2019''', '''\u2019''', ''')'''] if line.startswith('''@highlight''' ): return line if line[-1] in END_TOKENS: return line return line + "." def _A ( A__ , A__ , A__ ): """simple docstring""" if len(A__ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(A__ )) ) return sequence def _A ( A__ , A__ ): """simple docstring""" __lowercase = torch.ones_like(A__ ) __lowercase = sequence == pad_token_id __lowercase = 0 return mask def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = [tokenizer.encode(A__ ) for line in story_lines] __lowercase = [token for sentence in story_lines_token_ids for token in sentence] __lowercase = [tokenizer.encode(A__ ) for line in summary_lines] __lowercase = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] for sequence in batch: __lowercase = -1 __lowercase = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(A__ ) return torch.tensor(A__ )
624
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
1
'''simple docstring''' import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def _A ( A__ , A__=10 ): """simple docstring""" __lowercase = [] for _ in range(A__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def _A ( A__ , A__=10 ): """simple docstring""" __lowercase = [] for step in range(A__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(A__ , '''schedule.bin''' ) torch.save(scheduler.state_dict() , A__ ) __lowercase = torch.load(A__ ) scheduler.load_state_dict(A__ ) return lrs @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ): self.assertEqual(len(lowercase__ ) ,len(lowercase__ ) ) for a, b in zip(lowercase__ ,lowercase__ ): self.assertAlmostEqual(lowercase__ ,lowercase__ ,delta=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = torch.tensor([0.1, -0.2, -0.1] ,requires_grad=lowercase__ ) __lowercase = torch.tensor([0.4, 0.2, -0.5] ) __lowercase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping __lowercase = AdamW(params=[w] ,lr=2e-1 ,weight_decay=0.0 ) for _ in range(1_0_0 ): __lowercase = criterion(lowercase__ ,lowercase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() ,[0.4, 0.2, -0.5] ,tol=1e-2 ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = torch.tensor([0.1, -0.2, -0.1] ,requires_grad=lowercase__ ) __lowercase = torch.tensor([0.4, 0.2, -0.5] ) __lowercase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping __lowercase = Adafactor( params=[w] ,lr=1e-2 ,eps=(1e-3_0, 1e-3) ,clip_threshold=1.0 ,decay_rate=-0.8 ,betaa=lowercase__ ,weight_decay=0.0 ,relative_step=lowercase__ ,scale_parameter=lowercase__ ,warmup_init=lowercase__ ,) for _ in range(1_0_0_0 ): __lowercase = criterion(lowercase__ ,lowercase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() ,[0.4, 0.2, -0.5] ,tol=1e-2 ) @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = nn.Linear(5_0 , 5_0 ) if is_torch_available() else None SCREAMING_SNAKE_CASE : List[str] = AdamW(m.parameters() , lr=10.0 ) if is_torch_available() else None SCREAMING_SNAKE_CASE : Tuple = 1_0 def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Tuple ,lowercase__ : int ,lowercase__ : List[Any] ,lowercase__ : Optional[int]=None ): self.assertEqual(len(lowercase__ ) ,len(lowercase__ ) ) for a, b in zip(lowercase__ ,lowercase__ ): self.assertAlmostEqual(lowercase__ ,lowercase__ ,delta=lowercase__ ,msg=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = {'''num_warmup_steps''': 2, '''num_training_steps''': 1_0} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) __lowercase = { get_constant_schedule: ({}, [1_0.0] * self.num_steps), get_constant_schedule_with_warmup: ( {'''num_warmup_steps''': 4}, [0.0, 2.5, 5.0, 7.5, 1_0.0, 1_0.0, 1_0.0, 1_0.0, 1_0.0, 1_0.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 1_0.0, 8.7_5, 7.5, 6.2_5, 5.0, 3.7_5, 2.5, 1.2_5], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 1_0.0, 9.6_1, 8.5_3, 6.9_1, 5.0, 3.0_8, 1.4_6, 0.3_8], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, '''num_cycles''': 2}, [0.0, 5.0, 1_0.0, 8.5_3, 5.0, 1.4_6, 1_0.0, 8.5_3, 5.0, 1.4_6], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, '''power''': 2.0, '''lr_end''': 1e-7}, [0.0, 5.0, 1_0.0, 7.6_5_6, 5.6_2_5, 3.9_0_6, 2.5, 1.4_0_6, 0.6_2_5, 0.1_5_6], ), get_inverse_sqrt_schedule: ( {'''num_warmup_steps''': 2}, [0.0, 5.0, 1_0.0, 8.1_6_5, 7.0_7_1, 6.3_2_5, 5.7_7_4, 5.3_4_5, 5.0, 4.7_1_4], ), } for scheduler_func, data in scheds.items(): __lowercase , __lowercase = data __lowercase = scheduler_func(self.optimizer ,**lowercase__ ) self.assertEqual(len([scheduler.get_lr()[0]] ) ,1 ) __lowercase = unwrap_schedule(lowercase__ ,self.num_steps ) self.assertListAlmostEqual( lowercase__ ,lowercase__ ,tol=1e-2 ,msg=F"failed for {scheduler_func} in normal scheduler" ,) __lowercase = scheduler_func(self.optimizer ,**lowercase__ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(lowercase__ ) # wrap to test picklability of the schedule __lowercase = unwrap_and_save_reload_schedule(lowercase__ ,self.num_steps ) self.assertListEqual(lowercase__ ,lowercase__ ,msg=F"failed for {scheduler_func} in save and reload" ) class lowercase_ : """simple docstring""" def __init__( self : Tuple ,lowercase__ : Optional[Any] ): __lowercase = fn def __call__( self : Union[str, Any] ,*lowercase__ : Union[str, Any] ,**lowercase__ : List[Any] ): return self.fn(*lowercase__ ,**lowercase__ ) @classmethod def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Dict ): __lowercase = list(map(self ,scheduler.lr_lambdas ) )
624
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
1
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def _A ( A__ , A__ , A__ , A__ = 100 , ): """simple docstring""" __lowercase = x_start __lowercase = fnc(A__ ) __lowercase = 0.0 for _ in range(A__ ): # Approximates curve as a sequence of linear lines and sums their length __lowercase = (x_end - x_start) / steps + xa __lowercase = fnc(A__ ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step __lowercase = xa __lowercase = fxa return length if __name__ == "__main__": def _A ( A__ ): """simple docstring""" return math.sin(10 * x ) print('''f(x) = sin(10 * x)''') print('''The length of the curve from x = -10 to x = 10 is:''') lowerCAmelCase__ = 10 while i <= 10_0000: print(f'With {i} steps: {line_length(f, -10, 10, i)}') i *= 10
624
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
1
'''simple docstring''' import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() lowerCAmelCase__ = 2 class lowercase_ : """simple docstring""" def __init__( self : str ,*, # begin keyword-only arguments lowercase__ : Optional[int]="<s>" ,lowercase__ : Any="<pad>" ,lowercase__ : str="</s>" ,lowercase__ : Optional[Any]="<unk>" ,lowercase__ : Optional[Any]=None ,): __lowercase , __lowercase , __lowercase , __lowercase = bos, unk, pad, eos __lowercase = [] __lowercase = [] __lowercase = {} __lowercase = self.add_symbol(lowercase__ ) __lowercase = self.add_symbol(lowercase__ ) __lowercase = self.add_symbol(lowercase__ ) __lowercase = self.add_symbol(lowercase__ ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(lowercase__ ) __lowercase = len(self.symbols ) def __eq__( self : Dict ,lowercase__ : int ): return self.indices == other.indices def __getitem__( self : str ,lowercase__ : List[Any] ): if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self : List[str] ): return len(self.symbols ) def __contains__( self : List[str] ,lowercase__ : Optional[int] ): return sym in self.indices @classmethod def SCREAMING_SNAKE_CASE ( cls : str ,lowercase__ : Tuple ): __lowercase = cls() d.add_from_file(lowercase__ ) return d def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[str] ,lowercase__ : Optional[int]=1 ,lowercase__ : Dict=False ): if word in self.indices and not overwrite: __lowercase = self.indices[word] __lowercase = self.count[idx] + n return idx else: __lowercase = len(self.symbols ) __lowercase = idx self.symbols.append(lowercase__ ) self.count.append(lowercase__ ) return idx def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ): return 0 def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : int ): if isinstance(lowercase__ ,lowercase__ ): try: with open(lowercase__ ,'''r''' ,encoding='''utf-8''' ) as fd: self.add_from_file(lowercase__ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('''Incorrect encoding detected in {}, please rebuild the dataset'''.format(lowercase__ ) ) return __lowercase = f.readlines() __lowercase = self._load_meta(lowercase__ ) for line in lines[indices_start_line:]: try: __lowercase , __lowercase = line.rstrip().rsplit(''' ''' ,1 ) if field == "#fairseq:overwrite": __lowercase = True __lowercase , __lowercase = line.rsplit(''' ''' ,1 ) else: __lowercase = False __lowercase = int(lowercase__ ) __lowercase = line if word in self and not overwrite: raise RuntimeError( '''Duplicate word found when loading Dictionary: \'{}\'. ''' '''Duplicate words can overwrite earlier ones by adding the ''' '''#fairseq:overwrite flag at the end of the corresponding row ''' '''in the dictionary file. If using the Camembert model, please ''' '''download an updated copy of the model file.'''.format(lowercase__ ) ) self.add_symbol(lowercase__ ,n=lowercase__ ,overwrite=lowercase__ ) except ValueError: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt> [flags]\'''' ) def _A ( A__ ): """simple docstring""" __lowercase = dict((re.sub(R'''@@$''' , '''''' , A__ ), v) if k.endswith('''@@''' ) else (re.sub(R'''$''' , '''</w>''' , A__ ), v) for k, v in d.items() ) __lowercase = '''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[F"{k}</w>"] __lowercase = d[k] # restore return da def _A ( A__ , A__ ): """simple docstring""" if not os.path.exists(A__ ): raise ValueError(F"path {biogpt_checkpoint_path} does not exist!" ) os.makedirs(A__ , exist_ok=A__ ) print(F"Writing results to {pytorch_dump_folder_path}" ) # handle various types of models __lowercase = os.path.join(A__ , '''checkpoint.pt''' ) if not os.path.isfile(A__ ): raise ValueError(F"path to the file {checkpoint_file} does not exist!" ) __lowercase = torch.load(A__ , map_location='''cpu''' ) __lowercase = chkpt['''cfg''']['''model'''] # dicts __lowercase = os.path.join(A__ , '''dict.txt''' ) if not os.path.isfile(A__ ): raise ValueError(F"path to the file {dict_file} does not exist!" ) __lowercase = Dictionary.load(A__ ) __lowercase = rewrite_dict_keys(src_dict.indices ) __lowercase = len(A__ ) __lowercase = os.path.join(A__ , VOCAB_FILES_NAMES['''vocab_file'''] ) print(F"Generating {src_vocab_file} of {src_vocab_size} records" ) with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(A__ , ensure_ascii=A__ , indent=A__ ) ) # merges_file (bpecodes) __lowercase = os.path.join(A__ , '''bpecodes''' ) if not os.path.isfile(A__ ): raise ValueError(F"path to the file {bpecodes_file} does not exist!" ) __lowercase = os.path.join(A__ , VOCAB_FILES_NAMES['''merges_file'''] ) shutil.copyfile(A__ , A__ ) # model config __lowercase = os.path.join(A__ , '''config.json''' ) __lowercase = { '''activation_dropout''': args['''activation_dropout'''], '''architectures''': ['''BioGptForCausalLM'''], '''attention_probs_dropout_prob''': args['''attention_dropout'''], '''bos_token_id''': 0, '''eos_token_id''': 2, '''hidden_act''': args['''activation_fn'''], '''hidden_dropout_prob''': args['''dropout'''], '''hidden_size''': args['''decoder_embed_dim'''], '''initializer_range''': 0.0_2, '''intermediate_size''': args['''decoder_ffn_embed_dim'''], '''layer_norm_eps''': 1e-12, '''layerdrop''': args['''decoder_layerdrop'''], '''max_position_embeddings''': args['''max_target_positions'''], '''model_type''': '''biogpt''', '''num_attention_heads''': args['''decoder_attention_heads'''], '''num_hidden_layers''': args['''decoder_layers'''], '''pad_token_id''': 1, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_decoder_input_output_embed'''], '''vocab_size''': src_vocab_size, } # good hparam defaults to start with print(F"Generating {biogpt_model_config_file}" ) with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(A__ , ensure_ascii=A__ , indent=A__ ) ) # tokenizer config __lowercase = os.path.join(A__ , A__ ) __lowercase = { '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''model_max_length''': 1024, '''pad_token''': '''<pad>''', '''special_tokens_map_file''': None, '''tokenizer_class''': '''BioGptTokenizer''', '''unk_token''': '''<unk>''', } print(F"Generating {biogpt_tokenizer_config_file}" ) with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(A__ , ensure_ascii=A__ , indent=A__ ) ) # model __lowercase = chkpt['''model'''] # remove unneeded keys __lowercase = [ '''decoder.version''', ] for k in ignore_keys: model_state_dict.pop(A__ , A__ ) __lowercase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('''output_projection.weight''' ): __lowercase = model_state_dict.pop(A__ ) else: __lowercase = model_state_dict.pop(A__ ) __lowercase = BioGptConfig.from_pretrained(A__ ) __lowercase = BioGptForCausalLM(A__ ) # check that it loads ok model_new.load_state_dict(A__ ) # save __lowercase = os.path.join(A__ , A__ ) print(F"Generating {pytorch_weights_dump_path}" ) torch.save(A__ , A__ ) print('''Conversion is done!''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
624
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
1
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" if not (isinstance(A__ , A__ ) and isinstance(A__ , A__ )): raise ValueError('''longest_common_substring() takes two strings for inputs''' ) __lowercase = len(A__ ) __lowercase = len(A__ ) __lowercase = [[0] * (texta_length + 1) for _ in range(texta_length + 1 )] __lowercase = 0 __lowercase = 0 for i in range(1 , texta_length + 1 ): for j in range(1 , texta_length + 1 ): if texta[i - 1] == texta[j - 1]: __lowercase = 1 + dp[i - 1][j - 1] if dp[i][j] > ans_length: __lowercase = i __lowercase = dp[i][j] return texta[ans_index - ans_length : ans_index] if __name__ == "__main__": import doctest doctest.testmod()
624
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ ): """simple docstring""" __lowercase = SwinConfig( embed_dim=192 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=['''stage2''', '''stage3''', '''stage4'''] , ) __lowercase = DetaConfig( backbone_config=A__ , num_queries=900 , encoder_ffn_dim=2048 , decoder_ffn_dim=2048 , num_feature_levels=5 , assign_first_stage=A__ , with_box_refine=A__ , two_stage=A__ , ) # set labels __lowercase = '''huggingface/label-files''' if "o365" in model_name: __lowercase = 366 __lowercase = '''object365-id2label.json''' else: __lowercase = 91 __lowercase = '''coco-detection-id2label.json''' __lowercase = num_labels __lowercase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type='''dataset''' ) ) , '''r''' ) ) __lowercase = {int(A__ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} return config def _A ( A__ ): """simple docstring""" __lowercase = [] # stem # fmt: off rename_keys.append(('''backbone.0.body.patch_embed.proj.weight''', '''model.backbone.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.0.body.patch_embed.proj.bias''', '''model.backbone.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.0.body.patch_embed.norm.weight''', '''model.backbone.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.0.body.patch_embed.norm.bias''', '''model.backbone.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm1.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm1.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm2.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm2.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias") ) if i < 3: rename_keys.append((F"backbone.0.body.layers.{i}.downsample.reduction.weight", F"model.backbone.model.encoder.layers.{i}.downsample.reduction.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.downsample.norm.weight", F"model.backbone.model.encoder.layers.{i}.downsample.norm.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.downsample.norm.bias", F"model.backbone.model.encoder.layers.{i}.downsample.norm.bias") ) rename_keys.append(('''backbone.0.body.norm1.weight''', '''model.backbone.model.hidden_states_norms.stage2.weight''') ) rename_keys.append(('''backbone.0.body.norm1.bias''', '''model.backbone.model.hidden_states_norms.stage2.bias''') ) rename_keys.append(('''backbone.0.body.norm2.weight''', '''model.backbone.model.hidden_states_norms.stage3.weight''') ) rename_keys.append(('''backbone.0.body.norm2.bias''', '''model.backbone.model.hidden_states_norms.stage3.bias''') ) rename_keys.append(('''backbone.0.body.norm3.weight''', '''model.backbone.model.hidden_states_norms.stage4.weight''') ) rename_keys.append(('''backbone.0.body.norm3.bias''', '''model.backbone.model.hidden_states_norms.stage4.bias''') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight", F"model.encoder.layers.{i}.self_attn.sampling_offsets.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias", F"model.encoder.layers.{i}.self_attn.sampling_offsets.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.attention_weights.weight", F"model.encoder.layers.{i}.self_attn.attention_weights.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.attention_weights.bias", F"model.encoder.layers.{i}.self_attn.attention_weights.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.value_proj.weight", F"model.encoder.layers.{i}.self_attn.value_proj.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.value_proj.bias", F"model.encoder.layers.{i}.self_attn.value_proj.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.output_proj.weight", F"model.encoder.layers.{i}.self_attn.output_proj.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.output_proj.bias", F"model.encoder.layers.{i}.self_attn.output_proj.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm1.weight", F"model.encoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm1.bias", F"model.encoder.layers.{i}.self_attn_layer_norm.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear1.weight", F"model.encoder.layers.{i}.fc1.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear1.bias", F"model.encoder.layers.{i}.fc1.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear2.weight", F"model.encoder.layers.{i}.fc2.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear2.bias", F"model.encoder.layers.{i}.fc2.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm2.weight", F"model.encoder.layers.{i}.final_layer_norm.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm2.bias", F"model.encoder.layers.{i}.final_layer_norm.bias") ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight", F"model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias", F"model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.attention_weights.weight", F"model.decoder.layers.{i}.encoder_attn.attention_weights.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.attention_weights.bias", F"model.decoder.layers.{i}.encoder_attn.attention_weights.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.value_proj.weight", F"model.decoder.layers.{i}.encoder_attn.value_proj.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.value_proj.bias", F"model.decoder.layers.{i}.encoder_attn.value_proj.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.output_proj.weight", F"model.decoder.layers.{i}.encoder_attn.output_proj.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.output_proj.bias", F"model.decoder.layers.{i}.encoder_attn.output_proj.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm1.weight", F"model.decoder.layers.{i}.encoder_attn_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm1.bias", F"model.decoder.layers.{i}.encoder_attn_layer_norm.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.self_attn.out_proj.weight", F"model.decoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.self_attn.out_proj.bias", F"model.decoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm2.weight", F"model.decoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm2.bias", F"model.decoder.layers.{i}.self_attn_layer_norm.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear1.weight", F"model.decoder.layers.{i}.fc1.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear1.bias", F"model.decoder.layers.{i}.fc1.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear2.weight", F"model.decoder.layers.{i}.fc2.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear2.bias", F"model.decoder.layers.{i}.fc2.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm3.weight", F"model.decoder.layers.{i}.final_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm3.bias", F"model.decoder.layers.{i}.final_layer_norm.bias") ) # fmt: on return rename_keys def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = dct.pop(A__ ) __lowercase = val def _A ( A__ , A__ ): """simple docstring""" __lowercase = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __lowercase = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __lowercase = state_dict.pop(F"backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight" ) __lowercase = state_dict.pop(F"backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase = in_proj_weight[:dim, :] __lowercase = in_proj_bias[: dim] __lowercase = in_proj_weight[ dim : dim * 2, : ] __lowercase = in_proj_bias[ dim : dim * 2 ] __lowercase = in_proj_weight[ -dim :, : ] __lowercase = in_proj_bias[-dim :] # fmt: on def _A ( A__ , A__ ): """simple docstring""" __lowercase = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention __lowercase = state_dict.pop(F"transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) __lowercase = state_dict.pop(F"transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase = in_proj_weight[:hidden_size, :] __lowercase = in_proj_bias[:hidden_size] __lowercase = in_proj_weight[ hidden_size : hidden_size * 2, : ] __lowercase = in_proj_bias[hidden_size : hidden_size * 2] __lowercase = in_proj_weight[-hidden_size:, :] __lowercase = in_proj_bias[-hidden_size:] def _A ( ): """simple docstring""" __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = get_deta_config(A__ ) # load original state dict if model_name == "deta-swin-large": __lowercase = hf_hub_download(repo_id='''nielsr/deta-checkpoints''' , filename='''adet_swin_ft.pth''' ) elif model_name == "deta-swin-large-o365": __lowercase = hf_hub_download(repo_id='''jozhang97/deta-swin-l-o365''' , filename='''deta_swin_pt_o365.pth''' ) else: raise ValueError(F"Model name {model_name} not supported" ) __lowercase = torch.load(A__ , map_location='''cpu''' )['''model'''] # original state dict for name, param in state_dict.items(): print(A__ , param.shape ) # rename keys __lowercase = create_rename_keys(A__ ) for src, dest in rename_keys: rename_key(A__ , A__ , A__ ) read_in_swin_q_k_v(A__ , config.backbone_config ) read_in_decoder_q_k_v(A__ , A__ ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: __lowercase = state_dict.pop(A__ ) __lowercase = val if "input_proj" in key: __lowercase = state_dict.pop(A__ ) __lowercase = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: __lowercase = state_dict.pop(A__ ) __lowercase = val # finally, create HuggingFace model and load state dict __lowercase = DetaForObjectDetection(A__ ) model.load_state_dict(A__ ) model.eval() __lowercase = '''cuda''' if torch.cuda.is_available() else '''cpu''' model.to(A__ ) # load image processor __lowercase = DetaImageProcessor(format='''coco_detection''' ) # verify our conversion on image __lowercase = prepare_img() __lowercase = processor(images=A__ , return_tensors='''pt''' ) __lowercase = encoding['''pixel_values'''] __lowercase = model(pixel_values.to(A__ ) ) # verify logits print('''Logits:''' , outputs.logits[0, :3, :3] ) print('''Boxes:''' , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": __lowercase = torch.tensor( [[-7.6_3_0_8, -2.8_4_8_5, -5.3_7_3_7], [-7.2_0_3_7, -4.5_5_0_5, -4.8_0_2_7], [-7.2_9_4_3, -4.2_6_1_1, -4.6_6_1_7]] ) __lowercase = torch.tensor([[0.4_9_8_7, 0.4_9_6_9, 0.9_9_9_9], [0.2_5_4_9, 0.5_4_9_8, 0.4_8_0_5], [0.5_4_9_8, 0.2_7_5_7, 0.0_5_6_9]] ) elif model_name == "deta-swin-large-o365": __lowercase = torch.tensor( [[-8.0_1_2_2, -3.5_7_2_0, -4.9_7_1_7], [-8.1_5_4_7, -3.6_8_8_6, -4.6_3_8_9], [-7.6_6_1_0, -3.6_1_9_4, -5.0_1_3_4]] ) __lowercase = torch.tensor([[0.2_5_2_3, 0.5_5_4_9, 0.4_8_8_1], [0.7_7_1_5, 0.4_1_4_9, 0.4_6_0_1], [0.5_5_0_3, 0.2_7_5_3, 0.0_5_7_5]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(A__ ) , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(A__ ) , atol=1e-4 ) print('''Everything ok!''' ) if pytorch_dump_folder_path: # Save model and processor logger.info(F"Saving PyTorch model and processor to {pytorch_dump_folder_path}..." ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) processor.save_pretrained(A__ ) # Push to hub if push_to_hub: print('''Pushing model and processor to hub...''' ) model.push_to_hub(F"jozhang97/{model_name}" ) processor.push_to_hub(F"jozhang97/{model_name}" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', type=str, default='''deta-swin-large''', choices=['''deta-swin-large''', '''deta-swin-large-o365'''], help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCAmelCase__ = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
624
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''tiiuae/falcon-40b''': '''https://huggingface.co/tiiuae/falcon-40b/resolve/main/config.json''', '''tiiuae/falcon-7b''': '''https://huggingface.co/tiiuae/falcon-7b/resolve/main/config.json''', } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = 'falcon' SCREAMING_SNAKE_CASE : Optional[int] = ['past_key_values'] def __init__( self : Optional[Any] ,lowercase__ : Any=6_5_0_2_4 ,lowercase__ : Dict=4_5_4_4 ,lowercase__ : str=3_2 ,lowercase__ : Dict=7_1 ,lowercase__ : Any=1e-5 ,lowercase__ : Tuple=0.0_2 ,lowercase__ : str=True ,lowercase__ : str=0.0 ,lowercase__ : List[Any]=0.0 ,lowercase__ : List[Any]=None ,lowercase__ : int=False ,lowercase__ : Optional[int]=False ,lowercase__ : Any=True ,lowercase__ : Any=True ,lowercase__ : Optional[Any]=False ,lowercase__ : int=1_1 ,lowercase__ : Any=1_1 ,**lowercase__ : int ,): __lowercase = vocab_size # Backward compatibility with n_embed kwarg __lowercase = kwargs.pop('''n_embed''' ,lowercase__ ) __lowercase = hidden_size if n_embed is None else n_embed __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = layer_norm_epsilon __lowercase = initializer_range __lowercase = use_cache __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = bos_token_id __lowercase = eos_token_id __lowercase = num_attention_heads if num_kv_heads is None else num_kv_heads __lowercase = alibi __lowercase = new_decoder_architecture __lowercase = multi_query # Ignored when new_decoder_architecture is True __lowercase = parallel_attn __lowercase = bias super().__init__(bos_token_id=lowercase__ ,eos_token_id=lowercase__ ,**lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.hidden_size // self.num_attention_heads @property def SCREAMING_SNAKE_CASE ( self : str ): return not self.alibi
624
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
1
'''simple docstring''' import warnings warnings.warn( '''memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: ''' '''`from accelerate import find_executable_batch_size` to avoid this warning.''', FutureWarning, )
624
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=7 ,lowercase__ : Optional[Any]=3 ,lowercase__ : Optional[int]=1_8 ,lowercase__ : str=3_0 ,lowercase__ : Any=4_0_0 ,lowercase__ : List[Any]=True ,lowercase__ : Union[str, Any]=3_2 ,lowercase__ : int=True ,): __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size_divisor __lowercase = do_rescale def SCREAMING_SNAKE_CASE ( self : List[Any] ): return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = GLPNImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = GLPNImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : Dict ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase__ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''size_divisor''' ) ) self.assertTrue(hasattr(lowercase__ ,'''resample''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_rescale''' ) ) def SCREAMING_SNAKE_CASE ( self : int ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,numpify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def SCREAMING_SNAKE_CASE ( self : Any ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,torchify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
624
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
1
'''simple docstring''' # Algorithm for the pigeonhole sorting def _A ( A__ ): """simple docstring""" __lowercase = min(A__ ) # min() finds the minimum value __lowercase = max(A__ ) # max() finds the maximum value __lowercase = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size __lowercase = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(A__ , A__ ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. __lowercase = 0 for count in range(A__ ): while holes[count] > 0: holes[count] -= 1 __lowercase = count + min_val i += 1 def _A ( ): """simple docstring""" __lowercase = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(A__ ) print('''Sorted order is:''' , ''' '''.join(A__ ) ) if __name__ == "__main__": main()
624
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
1
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowercase_ : """simple docstring""" def __init__( self : Tuple ,lowercase__ : Union[str, Any] ,lowercase__ : str=1_3 ,lowercase__ : Optional[int]=3_0 ,lowercase__ : List[Any]=2 ,lowercase__ : Optional[int]=3 ,lowercase__ : int=True ,lowercase__ : List[Any]=True ,lowercase__ : int=3_2 ,lowercase__ : Optional[Any]=5 ,lowercase__ : str=4 ,lowercase__ : Tuple=3_7 ,lowercase__ : Dict="gelu" ,lowercase__ : str=0.1 ,lowercase__ : str=0.1 ,lowercase__ : Optional[Any]=1_0 ,lowercase__ : Tuple=0.0_2 ,lowercase__ : Optional[int]=3 ,lowercase__ : Optional[int]=None ,lowercase__ : Tuple=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 2 def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return DeiTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Union[str, Any] ): __lowercase = DeiTModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,lowercase__ : int ): __lowercase = DeiTForMaskedImageModeling(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual( result.reconstruction.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowercase = 1 __lowercase = DeiTForMaskedImageModeling(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.reconstruction.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = self.type_sequence_label_size __lowercase = DeiTForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowercase = 1 __lowercase = DeiTForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Any = ( { 'feature-extraction': DeiTModel, 'image-classification': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Union[str, Any] = False def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = DeiTModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : int ): self.config_tester.run_common_tests() @unittest.skip(reason='''DeiT does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : Optional[Any]=False ): __lowercase = super()._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def SCREAMING_SNAKE_CASE ( self : List[str] ): if not self.model_tester.is_training: return __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowercase__ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowercase = False __lowercase = True for model_class in self.all_model_classes: if model_class in get_values(lowercase__ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue __lowercase = model_class(lowercase__ ) model.gradient_checkpointing_enable() model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowercase__ ), *get_values(lowercase__ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): __lowercase = problem_type['''title'''] __lowercase = problem_type['''num_labels'''] __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) if problem_type["num_labels"] > 1: __lowercase = inputs['''labels'''].unsqueeze(1 ).repeat(1 ,problem_type['''num_labels'''] ) __lowercase = inputs['''labels'''].to(problem_type['''dtype'''] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowercase__ ) as warning_list: __lowercase = model(**lowercase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def SCREAMING_SNAKE_CASE ( self : Any ): for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DeiTModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Tuple ): return ( DeiTImageProcessor.from_pretrained('''facebook/deit-base-distilled-patch16-224''' ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = DeiTForImageClassificationWithTeacher.from_pretrained('''facebook/deit-base-distilled-patch16-224''' ).to( lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([-1.0_2_6_6, 0.1_9_1_2, -1.2_8_6_1] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = DeiTModel.from_pretrained( '''facebook/deit-base-distilled-patch16-224''' ,torch_dtype=torch.floataa ,device_map='''auto''' ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ) __lowercase = inputs.pixel_values.to(lowercase__ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __lowercase = model(lowercase__ )
624
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
1
'''simple docstring''' import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): lowerCAmelCase__ = yaml.safe_load( '''\ name: "" allow_empty: false allow_empty_text: true subsections: - name: "Dataset Card for X" # First-level markdown heading allow_empty: false allow_empty_text: true subsections: - name: "Table of Contents" allow_empty: false allow_empty_text: false subsections: null - name: "Dataset Description" allow_empty: false allow_empty_text: false subsections: - name: "Dataset Summary" allow_empty: false allow_empty_text: false subsections: null - name: "Supported Tasks and Leaderboards" allow_empty: true allow_empty_text: true subsections: null - name: Languages allow_empty: false allow_empty_text: true subsections: null ''' ) lowerCAmelCase__ = { '''name''': '''root''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [ { '''name''': '''Dataset Card for My Dataset''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [ {'''name''': '''Table of Contents''', '''text''': '''Some text here.''', '''is_empty_text''': False, '''subsections''': []}, { '''name''': '''Dataset Description''', '''text''': '''Some text here.''', '''is_empty_text''': False, '''subsections''': [ { '''name''': '''Dataset Summary''', '''text''': '''Some text here.''', '''is_empty_text''': False, '''subsections''': [], }, { '''name''': '''Supported Tasks and Leaderboards''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [], }, {'''name''': '''Languages''', '''text''': '''Language Text''', '''is_empty_text''': False, '''subsections''': []}, ], }, ], } ], } lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. #### Extra Ignored Subsection ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = { '''name''': '''root''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [ { '''name''': '''Dataset Card for My Dataset''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [ {'''name''': '''Table of Contents''', '''text''': '''Some text here.''', '''is_empty_text''': False, '''subsections''': []}, { '''name''': '''Dataset Description''', '''text''': '''Some text here.''', '''is_empty_text''': False, '''subsections''': [ { '''name''': '''Dataset Summary''', '''text''': '''Some text here.''', '''is_empty_text''': False, '''subsections''': [ { '''name''': '''Extra Ignored Subsection''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [], } ], }, { '''name''': '''Supported Tasks and Leaderboards''', '''text''': '''''', '''is_empty_text''': True, '''subsections''': [], }, {'''name''': '''Languages''', '''text''': '''Language Text''', '''is_empty_text''': False, '''subsections''': []}, ], }, ], } ], } lowerCAmelCase__ = '''\ --- --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = ( '''The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README.''' ) lowerCAmelCase__ = '''\ # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = ( '''The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README.''' ) lowerCAmelCase__ = '''\ --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored).''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found \'None\'.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Languages Language Text ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text # Dataset Card My Dataset ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README.''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README.''' lowerCAmelCase__ = '''\ --- language: - zh - en --- # Dataset Card for My Dataset # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text ''' lowerCAmelCase__ = '''The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections.''' @pytest.mark.parametrize( '''readme_md, expected_dict''' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def _A ( A__ , A__ ): """simple docstring""" assert ReadMe.from_string(A__ , A__ ).to_dict() == expected_dict @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def _A ( A__ , A__ ): """simple docstring""" with pytest.raises(A__ , match=re.escape(expected_error.format(path='''root''' ) ) ): __lowercase = ReadMe.from_string(A__ , A__ ) readme.validate() @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def _A ( A__ , A__ ): """simple docstring""" with pytest.raises(A__ , match=re.escape(expected_error.format(path='''root''' ) ) ): ReadMe.from_string(A__ , A__ ) @pytest.mark.parametrize( '''readme_md,''' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def _A ( A__ ): """simple docstring""" ReadMe.from_string(A__ , A__ , suppress_parsing_errors=A__ ) @pytest.mark.parametrize( '''readme_md, expected_dict''' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def _A ( A__ , A__ ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = Path(A__ ) / '''README.md''' with open(A__ , '''w+''' ) as readme_file: readme_file.write(A__ ) __lowercase = ReadMe.from_readme(A__ , A__ ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def _A ( A__ , A__ ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = Path(A__ ) / '''README.md''' with open(A__ , '''w+''' ) as readme_file: readme_file.write(A__ ) __lowercase = expected_error.format(path=A__ ) with pytest.raises(A__ , match=re.escape(A__ ) ): __lowercase = ReadMe.from_readme(A__ , A__ ) readme.validate() @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def _A ( A__ , A__ ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = Path(A__ ) / '''README.md''' with open(A__ , '''w+''' ) as readme_file: readme_file.write(A__ ) __lowercase = expected_error.format(path=A__ ) with pytest.raises(A__ , match=re.escape(A__ ) ): ReadMe.from_readme(A__ , A__ ) @pytest.mark.parametrize( '''readme_md,''' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def _A ( A__ ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = Path(A__ ) / '''README.md''' with open(A__ , '''w+''' ) as readme_file: readme_file.write(A__ ) ReadMe.from_readme(A__ , A__ , suppress_parsing_errors=A__ )
624
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
1
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ ): """simple docstring""" __lowercase = torch.load(A__ , map_location='''cpu''' ) if "model" in sd.keys(): __lowercase = torch.load(A__ , map_location='''cpu''' )['''model'''] # pop unnecessary weights __lowercase = [ '''decoder.version''', '''decoder.output_projection.weight''', ] for key in keys_to_delete: if key in sd: sd.pop(A__ ) __lowercase = { '''decoder.project_in_dim.weight''': '''decoder.project_in.weight''', '''decoder.project_out_dim.weight''': '''decoder.project_out.weight''', '''decoder.layer_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.layer_norm.bias''': '''decoder.final_layer_norm.bias''', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __lowercase = sd.pop(A__ ) __lowercase = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __lowercase = sd[key] # We split QKV in separate Q,K,V __lowercase = key.replace('''.qkv_proj.''' , '''.q_proj.''' ) __lowercase = key.replace('''.qkv_proj.''' , '''.k_proj.''' ) __lowercase = key.replace('''.qkv_proj.''' , '''.v_proj.''' ) __lowercase = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __lowercase , __lowercase , __lowercase = torch.split(A__ , depth // 3 , dim=0 ) __lowercase = q __lowercase = k __lowercase = v del sd[key] return sd @torch.no_grad() def _A ( A__ , A__ , A__=None ): """simple docstring""" __lowercase = load_checkpoint(A__ ) if config is not None: __lowercase = OPTConfig.from_pretrained(A__ ) else: __lowercase = OPTConfig() __lowercase = OPTModel(A__ ).half().eval() model.load_state_dict(A__ ) # Check results Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') lowerCAmelCase__ = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
624
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
1
'''simple docstring''' import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def _A ( A__ = 3 ): """simple docstring""" if isinstance(A__ , A__ ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(A__ ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) __lowercase = QuantumRegister(A__ , '''qr''' ) __lowercase = ClassicalRegister(A__ , '''cr''' ) __lowercase = QuantumCircuit(A__ , A__ ) __lowercase = number_of_qubits for i in range(A__ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(A__ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , A__ , A__ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(A__ , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(A__ , A__ ) # simulate with 10000 shots __lowercase = Aer.get_backend('''qasm_simulator''' ) __lowercase = execute(A__ , A__ , shots=10000 ) return job.result().get_counts(A__ ) if __name__ == "__main__": print( f'Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}' )
624
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
1
'''simple docstring''' import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''hidden_sizes''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_attention_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Any ,lowercase__ : Optional[int]=1_3 ,lowercase__ : Union[str, Any]=6_4 ,lowercase__ : List[str]=3 ,lowercase__ : Any=3 ,lowercase__ : Any=2 ,lowercase__ : Optional[int]=1 ,lowercase__ : Tuple=1_6 ,lowercase__ : List[str]=[1_2_8, 2_5_6, 3_8_4] ,lowercase__ : List[str]=[4, 6, 8] ,lowercase__ : Tuple=[2, 3, 4] ,lowercase__ : str=[1_6, 1_6, 1_6] ,lowercase__ : Optional[int]=0 ,lowercase__ : List[Any]=[2, 2, 2] ,lowercase__ : List[str]=[2, 2, 2] ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : List[str]=True ,lowercase__ : Optional[Any]=True ,lowercase__ : Any=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = kernel_size __lowercase = stride __lowercase = padding __lowercase = hidden_sizes __lowercase = num_attention_heads __lowercase = depths __lowercase = key_dim __lowercase = drop_path_rate __lowercase = patch_size __lowercase = attention_ratio __lowercase = mlp_ratio __lowercase = initializer_range __lowercase = [ ['''Subsample''', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['''Subsample''', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = initializer_range def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : List[Any] ): return LevitConfig( image_size=self.image_size ,num_channels=self.num_channels ,kernel_size=self.kernel_size ,stride=self.stride ,padding=self.padding ,patch_size=self.patch_size ,hidden_sizes=self.hidden_sizes ,num_attention_heads=self.num_attention_heads ,depths=self.depths ,key_dim=self.key_dim ,drop_path_rate=self.drop_path_rate ,mlp_ratio=self.mlp_ratio ,attention_ratio=self.attention_ratio ,initializer_range=self.initializer_range ,down_ops=self.down_ops ,) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ,lowercase__ : Any ): __lowercase = LevitModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for _ in range(4 ): __lowercase = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) __lowercase = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) ,) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : str ): __lowercase = self.num_labels __lowercase = LevitForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': LevitModel, 'image-classification': (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Tuple = False def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = LevitModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : Any ): return @unittest.skip(reason='''Levit does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass @unittest.skip(reason='''Levit does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass @unittest.skip(reason='''Levit does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): def check_hidden_states_output(lowercase__ : str ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowercase__ ) ,lowercase__ ) __lowercase = (self.model_tester.image_size, self.model_tester.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for _ in range(4 ): __lowercase = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) __lowercase = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[ height * width, self.model_tester.hidden_sizes[0], ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : Tuple ,lowercase__ : Tuple=False ): __lowercase = super()._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): if not self.model_tester.is_training: return __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowercase__ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : int ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowercase = False __lowercase = True for model_class in self.all_model_classes: if model_class in get_values(lowercase__ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue __lowercase = model_class(lowercase__ ) model.gradient_checkpointing_enable() model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowercase__ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): __lowercase = problem_type['''title'''] __lowercase = problem_type['''num_labels'''] __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) if problem_type["num_labels"] > 1: __lowercase = inputs['''labels'''].unsqueeze(1 ).repeat(1 ,problem_type['''num_labels'''] ) __lowercase = inputs['''labels'''].to(problem_type['''dtype'''] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowercase__ ) as warning_list: __lowercase = model(**lowercase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = LevitModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Any ): return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([1.0_4_4_8, -0.3_7_4_5, -1.8_3_1_7] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
1
'''simple docstring''' from math import factorial lowerCAmelCase__ = {str(digit): factorial(digit) for digit in range(10)} def _A ( A__ ): """simple docstring""" if not isinstance(A__ , A__ ): raise TypeError('''Parameter number must be int''' ) if number < 0: raise ValueError('''Parameter number must be greater than or equal to 0''' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(A__ ) ) def _A ( A__ = 60 , A__ = 1000000 ): """simple docstring""" if not isinstance(A__ , A__ ) or not isinstance(A__ , A__ ): raise TypeError('''Parameters chain_length and number_limit must be int''' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( '''Parameters chain_length and number_limit must be greater than 0''' ) # the counter for the chains with the exact desired length __lowercase = 0 # the cached sizes of the previous chains __lowercase = {} for start_chain_element in range(1 , A__ ): # The temporary set will contain the elements of the chain __lowercase = set() __lowercase = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. __lowercase = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(A__ ) chain_set_length += 1 __lowercase = digit_factorial_sum(A__ ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] __lowercase = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(f'{solution()}')
624
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
1
'''simple docstring''' def _A ( A__ = 3 , A__ = 7 , A__ = 1000000 ): """simple docstring""" __lowercase = 0 __lowercase = 1 for current_denominator in range(1 , limit + 1 ): __lowercase = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: __lowercase = current_numerator __lowercase = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=100_0000))
624
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
1
'''simple docstring''' # HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers lowerCAmelCase__ = float('''nan''') class lowercase_ : """simple docstring""" def __init__( self : int ,lowercase__ : List[Any] ): __lowercase = sys.stdout __lowercase = open(lowercase__ ,'''a''' ) def __getattr__( self : str ,lowercase__ : List[Any] ): return getattr(self.stdout ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[Any] ): self.stdout.write(lowercase__ ) # strip tqdm codes self.file.write(re.sub(r'''^.*\r''' ,'''''' ,lowercase__ ,0 ,re.M ) ) def _A ( A__=80 , A__=False ): """simple docstring""" __lowercase = [] # deal with critical env vars __lowercase = ['''CUDA_VISIBLE_DEVICES'''] for key in env_keys: __lowercase = os.environ.get(A__ , A__ ) if val is not None: cmd.append(F"{key}={val}" ) # python executable (not always needed if the script is executable) __lowercase = sys.executable if full_python_path else sys.executable.split('''/''' )[-1] cmd.append(A__ ) # now the normal args cmd += list(map(shlex.quote , sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes __lowercase = [] __lowercase = '''''' while len(A__ ) > 0: current_line += F"{cmd.pop(0 )} " if len(A__ ) == 0 or len(A__ ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(A__ ) __lowercase = '''''' return "\\\n".join(A__ ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = re.sub(R'''[\\\n]+''' , ''' ''' , args.base_cmd ) # remove --output_dir if any and set our own __lowercase = re.sub('''--output_dir\s+[^\s]+''' , '''''' , args.base_cmd ) args.base_cmd += F" --output_dir {output_dir}" # ensure we have --overwrite_output_dir __lowercase = re.sub('''--overwrite_output_dir\s+''' , '''''' , args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def _A ( A__ , A__ , A__ , A__ , A__ , A__ , A__ ): """simple docstring""" if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 , 100 ) for k in metric_keys} , **{target_metric_key: random.choice([nan, 1_0.3_1, 1_0_0.2, 5_5.6_6_6_6, 2_2_2.2_2_2_2_2_2_2_2] )} , ) __lowercase = subprocess.run(A__ , capture_output=A__ , text=A__ ) if verbose: print('''STDOUT''' , result.stdout ) print('''STDERR''' , result.stderr ) # save the streams __lowercase = variation.replace(''' ''' , '''-''' ) with open(Path(A__ ) / F"log.{prefix}.stdout.txt" , '''w''' ) as f: f.write(result.stdout ) with open(Path(A__ ) / F"log.{prefix}.stderr.txt" , '''w''' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('''failed''' ) return {target_metric_key: nan} with io.open(F"{output_dir}/all_results.json" , '''r''' , encoding='''utf-8''' ) as f: __lowercase = json.load(A__ ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def _A ( A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ): """simple docstring""" __lowercase = [] __lowercase = [] __lowercase = F"{id}: {variation:<{longest_variation_len}}" __lowercase = F"{preamble}: " __lowercase = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(A__ ) , desc=A__ , leave=A__ ): __lowercase = process_run_single( A__ , A__ , A__ , A__ , A__ , A__ , A__ ) __lowercase = single_run_metrics[target_metric_key] if not math.isnan(A__ ): metrics.append(A__ ) results.append(A__ ) outcome += "✓" else: outcome += "✘" __lowercase = F"\33[2K\r{outcome}" if len(A__ ) > 0: __lowercase = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} __lowercase = round(mean_metrics[target_metric_key] , 2 ) __lowercase = F"{outcome} {mean_target}" if len(A__ ) > 1: results_str += F" {tuple(round(A__ , 2 ) for x in results )}" print(A__ ) __lowercase = variation return mean_metrics else: print(A__ ) return {variation_key: variation, target_metric_key: nan} def _A ( ): """simple docstring""" __lowercase = torch.cuda.get_device_properties(torch.device('''cuda''' ) ) return F"\nDatetime : {datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S' )}\n\nSoftware:\ntransformers: {transformers.__version__}\ntorch : {torch.__version__}\ncuda : {torch.version.cuda}\npython : {platform.python_version()}\n\nHardware:\n{torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB\n" def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = pd.DataFrame(A__ ) __lowercase = '''variation''' __lowercase = '''diff_%''' __lowercase = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan __lowercase = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(A__ ): # as a fallback, use the minimal value as the sentinel __lowercase = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(A__ ): __lowercase = df.apply( lambda A__ : round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 , axis='''columns''' , ) # re-order columns __lowercase = [variation_key, target_metric_key, diff_key, *report_metric_keys] __lowercase = df.reindex(A__ , axis='''columns''' ) # reorder cols # capitalize __lowercase = df.rename(str.capitalize , axis='''columns''' ) # make the cols as narrow as possible __lowercase = df.rename(lambda A__ : c.replace('''_''' , '''<br>''' ) , axis='''columns''' ) __lowercase = df.rename(lambda A__ : c.replace('''_''' , '''\n''' ) , axis='''columns''' ) __lowercase = ['''''', '''Copy between the cut-here-lines and paste as is to github or a forum'''] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=A__ , floatfmt='''.2f''' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=A__ , floatfmt='''.2f''' )] print('''\n\n'''.join(A__ ) ) def _A ( ): """simple docstring""" __lowercase = argparse.ArgumentParser() parser.add_argument( '''--base-cmd''' , default=A__ , type=A__ , required=A__ , help='''Base cmd''' , ) parser.add_argument( '''--variations''' , default=A__ , type=A__ , nargs='''+''' , required=A__ , help='''Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'''' , ) parser.add_argument( '''--base-variation''' , default=A__ , type=A__ , help='''Baseline variation to compare to. if None the minimal target value will be used to compare against''' , ) parser.add_argument( '''--target-metric-key''' , default=A__ , type=A__ , required=A__ , help='''Target metric key in output_dir/all_results.json, e.g., train_samples_per_second''' , ) parser.add_argument( '''--report-metric-keys''' , default='''''' , type=A__ , help='''Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples''' , ) parser.add_argument( '''--repeat-times''' , default=1 , type=A__ , help='''How many times to re-run each variation - an average will be reported''' , ) parser.add_argument( '''--output_dir''' , default='''output_benchmark''' , type=A__ , help='''The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked''' , ) parser.add_argument( '''--verbose''' , default=A__ , action='''store_true''' , help='''Whether to show the outputs of each run or just the benchmark progress''' , ) __lowercase = parser.parse_args() __lowercase = args.output_dir Path(A__ ).mkdir(exist_ok=A__ ) __lowercase = get_base_command(A__ , A__ ) # split each dimension into its --foo variations __lowercase = [list(map(str.strip , re.split(R'''\|''' , A__ ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty __lowercase = list(map(str.strip , map(''' '''.join , itertools.product(*A__ ) ) ) ) __lowercase = max(len(A__ ) for x in variations ) # split wanted keys __lowercase = args.report_metric_keys.split() # capture prints into a log file for convenience __lowercase = F"benchmark-report-{datetime.datetime.now().strftime('%Y-%m-%d-%H-%M-%S' )}.txt" print(F"\nNote: each run's output is also logged under {output_dir}/log.*.std*.txt" ) print(F"and this script's output is also piped into {report_fn}" ) __lowercase = Tee(A__ ) print(F"\n*** Running {len(A__ )} benchmarks:" ) print(F"Base command: {' '.join(A__ )}" ) __lowercase = '''variation''' __lowercase = [] for id, variation in enumerate(tqdm(A__ , desc='''Total completion: ''' , leave=A__ ) ): __lowercase = base_cmd + variation.split() results.append( process_run( id + 1 , A__ , A__ , A__ , A__ , args.target_metric_key , A__ , args.repeat_times , A__ , args.verbose , ) ) process_results(A__ , args.target_metric_key , A__ , args.base_variation , A__ ) if __name__ == "__main__": main()
624
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
1
'''simple docstring''' # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowerCAmelCase__ = get_logger(__name__) class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = 'dummy_data' SCREAMING_SNAKE_CASE : Tuple = 'datasets' SCREAMING_SNAKE_CASE : Union[str, Any] = False def __init__( self : Optional[int] ,lowercase__ : str ,lowercase__ : str ,lowercase__ : Union[Version, str] ,lowercase__ : Optional[str] = None ,lowercase__ : bool = False ,lowercase__ : bool = True ,lowercase__ : Optional[List[Callable]] = None ,): __lowercase = 0 __lowercase = dataset_name __lowercase = cache_dir __lowercase = use_local_dummy_data __lowercase = config # download_callbacks take a single url as input __lowercase = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root __lowercase = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general __lowercase = str(lowercase__ ) # to be downloaded __lowercase = None __lowercase = None @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): if self._dummy_file is None: __lowercase = self.download_dummy_data() return self._dummy_file @property def SCREAMING_SNAKE_CASE ( self : int ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('''dummy''' ,self.config.name ,self.version_name ) # structure is dummy / version_name return os.path.join('''dummy''' ,self.version_name ) @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return os.path.join(self.dummy_data_folder ,'''dummy_data.zip''' ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) __lowercase = cached_path( lowercase__ ,cache_dir=self.cache_dir ,extract_compressed_file=lowercase__ ,force_extract=lowercase__ ) return os.path.join(lowercase__ ,self.dummy_file_name ) @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return os.path.join(self.datasets_scripts_dir ,self.dataset_name ,self.dummy_zip_file ) @property def SCREAMING_SNAKE_CASE ( self : int ): if self._bucket_url is None: __lowercase = hf_github_url(self.dataset_name ,self.dummy_zip_file.replace(os.sep ,'''/''' ) ) return self._bucket_url @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep ,'''/''' ).split('''/''' )[:-1] ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Dict ,*lowercase__ : List[Any] ): if self.load_existing_dummy_data: # dummy data is downloaded and tested __lowercase = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned __lowercase = self.dummy_file_name # special case when data_url is a dict if isinstance(lowercase__ ,lowercase__ ): return self.create_dummy_data_dict(lowercase__ ,lowercase__ ) elif isinstance(lowercase__ ,(list, tuple) ): return self.create_dummy_data_list(lowercase__ ,lowercase__ ) else: return self.create_dummy_data_single(lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Optional[int] ,*lowercase__ : List[Any] ): return self.download_and_extract(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[int] ): return self.download_and_extract(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Any ,*lowercase__ : Tuple ,**lowercase__ : int ): return path def SCREAMING_SNAKE_CASE ( self : int ): return {} def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : int ,lowercase__ : int ): __lowercase = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowercase__ ,lowercase__ ): for single_url in single_urls: download_callback(lowercase__ ) else: __lowercase = single_urls download_callback(lowercase__ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowercase__ ,lowercase__ ): __lowercase = [os.path.join(lowercase__ ,urllib.parse.quote_plus(Path(lowercase__ ).name ) ) for x in single_urls] else: __lowercase = single_urls __lowercase = os.path.join(lowercase__ ,urllib.parse.quote_plus(Path(lowercase__ ).name ) ) __lowercase = value # make sure that values are unique if all(isinstance(lowercase__ ,lowercase__ ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique __lowercase = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Optional[int] ,lowercase__ : Dict ): __lowercase = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one __lowercase = all(bool(re.findall('''[0-9]{3,}-of-[0-9]{3,}''' ,lowercase__ ) ) for url in data_url ) __lowercase = all( url.startswith('''https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed''' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): __lowercase = [data_url[0]] * len(lowercase__ ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowercase__ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __lowercase = os.path.join(lowercase__ ,urllib.parse.quote_plus(single_url.split('''/''' )[-1] ) ) dummy_data_list.append(lowercase__ ) return dummy_data_list def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Tuple ,lowercase__ : Dict ): for download_callback in self.download_callbacks: download_callback(lowercase__ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __lowercase = os.path.join(lowercase__ ,urllib.parse.quote_plus(data_url.split('''/''' )[-1] ) ) if os.path.exists(lowercase__ ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def SCREAMING_SNAKE_CASE ( self : Any ): pass def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ): def _iter_archive_members(lowercase__ : int ): # this preserves the order of the members inside the ZIP archive __lowercase = Path(self.dummy_file ).parent __lowercase = path.relative_to(lowercase__ ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: __lowercase = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowercase__ ) __lowercase = Path(lowercase__ ) __lowercase = _iter_archive_members(lowercase__ ) if self.use_local_dummy_data else path.rglob('''*''' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('''.''', '''__''') ): yield file_path.relative_to(lowercase__ ).as_posix(), file_path.open('''rb''' ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : int ): if not isinstance(lowercase__ ,lowercase__ ): __lowercase = [paths] for path in paths: if os.path.isfile(lowercase__ ): if os.path.basename(lowercase__ ).startswith(('''.''', '''__''') ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowercase__ ): if os.path.basename(lowercase__ ).startswith(('''.''', '''__''') ): continue dirnames.sort() for filename in sorted(lowercase__ ): if filename.startswith(('''.''', '''__''') ): continue yield os.path.join(lowercase__ ,lowercase__ )
624
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
1
'''simple docstring''' from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-1''' lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-2''' lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-3''' lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-4''' class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[int] ,lowercase__ : AutoencoderKL ,lowercase__ : CLIPTextModel ,lowercase__ : CLIPTokenizer ,lowercase__ : UNetaDConditionModel ,lowercase__ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] ,lowercase__ : StableDiffusionSafetyChecker ,lowercase__ : CLIPImageProcessor ,lowercase__ : bool = True ,): super()._init_() __lowercase = StableDiffusionPipeline.from_pretrained(lowercase__ ) __lowercase = StableDiffusionPipeline.from_pretrained(lowercase__ ) __lowercase = StableDiffusionPipeline.from_pretrained(lowercase__ ) __lowercase = StableDiffusionPipeline( vae=lowercase__ ,text_encoder=lowercase__ ,tokenizer=lowercase__ ,unet=lowercase__ ,scheduler=lowercase__ ,safety_checker=lowercase__ ,feature_extractor=lowercase__ ,requires_safety_checker=lowercase__ ,) self.register_modules(pipelinea=self.pipea ,pipelinea=self.pipea ,pipelinea=self.pipea ,pipelinea=self.pipea ) @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return {k: getattr(self ,lowercase__ ) for k in self.config.keys() if not k.startswith('''_''' )} def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Optional[Union[str, int]] = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __lowercase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.enable_attention_slicing(lowercase__ ) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : int ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Optional[Any] ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Optional[int] ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Any ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : str ,): __lowercase = '''cuda''' if torch.cuda.is_available() else '''cpu''' self.to(lowercase__ ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` must be divisible by 8 but are {height} and {width}." ) # Get first result from Stable Diffusion Checkpoint v1.1 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get first result from Stable Diffusion Checkpoint v1.2 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get first result from Stable Diffusion Checkpoint v1.3 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get first result from Stable Diffusion Checkpoint v1.4 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
624
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
1
'''simple docstring''' import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def _A ( A__ ): """simple docstring""" if "model" in orig_key: __lowercase = orig_key.replace('''model.''' , '''''' ) if "norm1" in orig_key: __lowercase = orig_key.replace('''norm1''' , '''attention.output.LayerNorm''' ) if "norm2" in orig_key: __lowercase = orig_key.replace('''norm2''' , '''output.LayerNorm''' ) if "norm" in orig_key: __lowercase = orig_key.replace('''norm''' , '''LayerNorm''' ) if "transformer" in orig_key: __lowercase = orig_key.split('''.''' )[0].split('''_''' )[-1] __lowercase = orig_key.replace(F"transformer_{layer_num}" , F"encoder.layer.{layer_num}" ) if "mha.attn" in orig_key: __lowercase = orig_key.replace('''mha.attn''' , '''attention.self''' ) if "mha" in orig_key: __lowercase = orig_key.replace('''mha''' , '''attention''' ) if "W_q" in orig_key: __lowercase = orig_key.replace('''W_q''' , '''self.query''' ) if "W_k" in orig_key: __lowercase = orig_key.replace('''W_k''' , '''self.key''' ) if "W_v" in orig_key: __lowercase = orig_key.replace('''W_v''' , '''self.value''' ) if "ff1" in orig_key: __lowercase = orig_key.replace('''ff1''' , '''intermediate.dense''' ) if "ff2" in orig_key: __lowercase = orig_key.replace('''ff2''' , '''output.dense''' ) if "ff" in orig_key: __lowercase = orig_key.replace('''ff''' , '''output.dense''' ) if "mlm_class" in orig_key: __lowercase = orig_key.replace('''mlm.mlm_class''' , '''cls.predictions.decoder''' ) if "mlm" in orig_key: __lowercase = orig_key.replace('''mlm''' , '''cls.predictions.transform''' ) if "cls" not in orig_key: __lowercase = '''yoso.''' + orig_key return orig_key def _A ( A__ , A__ ): """simple docstring""" for key in orig_state_dict.copy().keys(): __lowercase = orig_state_dict.pop(A__ ) if ("pooler" in key) or ("sen_class" in key): continue else: __lowercase = val __lowercase = orig_state_dict['''cls.predictions.decoder.bias'''] __lowercase = torch.arange(A__ ).expand((1, -1) ) + 2 return orig_state_dict def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = torch.load(A__ , map_location='''cpu''' )['''model_state_dict'''] __lowercase = YosoConfig.from_json_file(A__ ) __lowercase = YosoForMaskedLM(A__ ) __lowercase = convert_checkpoint_helper(config.max_position_embeddings , A__ ) print(model.load_state_dict(A__ ) ) model.eval() model.save_pretrained(A__ ) print(F"Checkpoint successfuly converted. Model saved at {pytorch_dump_path}" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--pytorch_model_path''', default=None, type=str, required=True, help='''Path to YOSO pytorch checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The json file for YOSO model config.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
624
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
1
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
1
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" return price * (1 + tax_rate) if __name__ == "__main__": print(f'{price_plus_tax(100, 0.25) = }') print(f'{price_plus_tax(125.50, 0.05) = }')
624
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = 'megatron-bert' def __init__( self : Tuple ,lowercase__ : str=2_9_0_5_6 ,lowercase__ : Dict=1_0_2_4 ,lowercase__ : Any=2_4 ,lowercase__ : Optional[int]=1_6 ,lowercase__ : Any=4_0_9_6 ,lowercase__ : int="gelu" ,lowercase__ : Optional[Any]=0.1 ,lowercase__ : List[Any]=0.1 ,lowercase__ : str=5_1_2 ,lowercase__ : Any=2 ,lowercase__ : Tuple=0.0_2 ,lowercase__ : Dict=1e-1_2 ,lowercase__ : Any=0 ,lowercase__ : List[Any]="absolute" ,lowercase__ : Optional[Any]=True ,**lowercase__ : Any ,): super().__init__(pad_token_id=lowercase__ ,**lowercase__ ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache
624
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
1