code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class lowercase_ (unittest.TestCase ): def __init__( self : Union[str, Any] ,lowercase__ : str ,lowercase__ : Optional[int]=7 ,lowercase__ : Tuple=3 ,lowercase__ : List[Any]=1_8 ,lowercase__ : Any=3_0 ,lowercase__ : Optional[int]=4_0_0 ,lowercase__ : List[Any]=True ,lowercase__ : List[str]=None ,lowercase__ : Tuple=True ,lowercase__ : Dict=[0.5, 0.5, 0.5] ,lowercase__ : Tuple=[0.5, 0.5, 0.5] ,): __lowercase = size if size is not None else {'''height''': 1_8, '''width''': 1_8} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = do_normalize __lowercase = image_mean __lowercase = image_std def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class lowercase_ (lowerCamelCase__ , unittest.TestCase ): SCREAMING_SNAKE_CASE : Union[str, Any] = DPTImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = DPTImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : Any ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase__ ,'''image_mean''' ) ) self.assertTrue(hasattr(lowercase__ ,'''image_std''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_normalize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''size''' ) ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''height''': 1_8, '''width''': 1_8} ) __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ) self.assertEqual(image_processor.size ,{'''height''': 4_2, '''width''': 4_2} ) def SCREAMING_SNAKE_CASE ( self : Dict ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,numpify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,torchify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,)
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' from math import isqrt def _A ( A__ ): """simple docstring""" __lowercase = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , A__ , A__ ): __lowercase = False return [i for i in range(2 , A__ ) if is_prime[i]] def _A ( A__ = 10**8 ): """simple docstring""" __lowercase = calculate_prime_numbers(max_number // 2 ) __lowercase = 0 __lowercase = 0 __lowercase = len(A__ ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(f'{solution() = }')
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' from __future__ import annotations from collections.abc import Sequence from typing import Literal def _A ( A__ , A__ ): """simple docstring""" __lowercase = list(A__ ) __lowercase = list(A__ ) __lowercase = 0 for i in range(len(A__ ) ): if lista[i] != lista[i]: count += 1 __lowercase = '''_''' if count > 1: return False else: return "".join(A__ ) def _A ( A__ ): """simple docstring""" __lowercase = [] while True: __lowercase = ['''$'''] * len(A__ ) __lowercase = [] for i in range(len(A__ ) ): for j in range(i + 1 , len(A__ ) ): __lowercase = compare_string(binary[i] , binary[j] ) if k is False: __lowercase = '''*''' __lowercase = '''*''' temp.append('''X''' ) for i in range(len(A__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(A__ ) == 0: return pi __lowercase = list(set(A__ ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] for minterm in minterms: __lowercase = '''''' for _ in range(A__ ): __lowercase = str(minterm % 2 ) + string minterm //= 2 temp.append(A__ ) return temp def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = list(A__ ) __lowercase = list(A__ ) __lowercase = 0 for i in range(len(A__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [0] * len(A__ ) for i in range(len(chart[0] ) ): __lowercase = 0 __lowercase = -1 for j in range(len(A__ ) ): if chart[j][i] == 1: count += 1 __lowercase = j if count == 1: __lowercase = 1 for i in range(len(A__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(A__ ) ): __lowercase = 0 temp.append(prime_implicants[i] ) while True: __lowercase = 0 __lowercase = -1 __lowercase = 0 for i in range(len(A__ ) ): __lowercase = chart[i].count(1 ) if count_n > max_n: __lowercase = count_n __lowercase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(A__ ) ): __lowercase = 0 def _A ( A__ , A__ ): """simple docstring""" __lowercase = [[0 for x in range(len(A__ ) )] for x in range(len(A__ ) )] for i in range(len(A__ ) ): __lowercase = prime_implicants[i].count('''_''' ) for j in range(len(A__ ) ): if is_for_table(prime_implicants[i] , binary[j] , A__ ): __lowercase = 1 return chart def _A ( ): """simple docstring""" __lowercase = int(input('''Enter the no. of variables\n''' ) ) __lowercase = [ float(A__ ) for x in input( '''Enter the decimal representation of Minterms \'Spaces Separated\'\n''' ).split() ] __lowercase = decimal_to_binary(A__ , A__ ) __lowercase = check(A__ ) print('''Prime Implicants are:''' ) print(A__ ) __lowercase = prime_implicant_chart(A__ , A__ ) __lowercase = selection(A__ , A__ ) print('''Essential Prime Implicants are:''' ) print(A__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : str ,lowercase__ : int ,lowercase__ : Union[str, Any]=1_3 ,lowercase__ : List[str]=7 ,lowercase__ : Any=True ,lowercase__ : Tuple=True ,lowercase__ : Tuple=False ,lowercase__ : Any=True ,lowercase__ : List[Any]=9_9 ,lowercase__ : int=3_2 ,lowercase__ : Dict=5 ,lowercase__ : Optional[int]=4 ,lowercase__ : List[str]=3_7 ,lowercase__ : Optional[int]="gelu" ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : List[str]=0.1 ,lowercase__ : Optional[Any]=5_1_2 ,lowercase__ : str=1_6 ,lowercase__ : Dict=2 ,lowercase__ : Union[str, Any]=0.0_2 ,lowercase__ : int=3 ,lowercase__ : Tuple=4 ,lowercase__ : Optional[Any]=None ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Tuple ): return DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : List[str] ): __lowercase = DistilBertModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Optional[Any] ): __lowercase = DistilBertForMaskedLM(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : int ): __lowercase = DistilBertForQuestionAnswering(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,start_positions=lowercase__ ,end_positions=lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int] ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ): __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Any ,lowercase__ : Optional[int] ,lowercase__ : Tuple ,lowercase__ : Any ,lowercase__ : int ,lowercase__ : Union[str, Any] ): __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Optional[Any] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : str ): __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : str = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE : Optional[Any] = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : int = True SCREAMING_SNAKE_CASE : Union[str, Any] = True SCREAMING_SNAKE_CASE : Union[str, Any] = True SCREAMING_SNAKE_CASE : Any = True def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,dim=3_7 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) @slow @require_torch_gpu def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = torch.jit.trace( lowercase__ ,(inputs_dict['''input_ids'''].to('''cpu''' ), inputs_dict['''attention_mask'''].to('''cpu''' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowercase__ ,os.path.join(lowercase__ ,'''traced_model.pt''' ) ) __lowercase = torch.jit.load(os.path.join(lowercase__ ,'''traced_model.pt''' ) ,map_location=lowercase__ ) loaded(inputs_dict['''input_ids'''].to(lowercase__ ) ,inputs_dict['''attention_mask'''].to(lowercase__ ) ) @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = DistilBertModel.from_pretrained('''distilbert-base-uncased''' ) __lowercase = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(lowercase__ ,attention_mask=lowercase__ )[0] __lowercase = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape ,lowercase__ ) __lowercase = torch.tensor( [[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] ,lowercase__ ,atol=1e-4 ) )
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" if len(A__ ) <= 1: return [tuple(A__ )] __lowercase = [] def generate(A__ , A__ ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , A__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even __lowercase , __lowercase = arr[k - 1], arr[i] else: # k is odd __lowercase , __lowercase = arr[k - 1], arr[0] generate(k - 1 , A__ ) generate(len(A__ ) , A__ ) return res if __name__ == "__main__": lowerCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor lowerCAmelCase__ = logging.get_logger(__name__) class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : List[str] ,*lowercase__ : Dict ,**lowercase__ : Union[str, Any] ): warnings.warn( '''The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use VideoMAEImageProcessor instead.''' ,lowercase__ ,) super().__init__(*lowercase__ ,**lowercase__ )
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' import math from collections.abc import Iterator from itertools import takewhile def _A ( A__ ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(A__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _A ( ): """simple docstring""" __lowercase = 2 while True: if is_prime(A__ ): yield num num += 1 def _A ( A__ = 2000000 ): """simple docstring""" return sum(takewhile(lambda A__ : x < n , prime_generator() ) ) if __name__ == "__main__": print(f'{solution() = }')
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import argparse from collections import defaultdict import yaml lowerCAmelCase__ = '''docs/source/en/_toctree.yml''' def _A ( A__ ): """simple docstring""" __lowercase = defaultdict(A__ ) __lowercase = [] __lowercase = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({'''local''': doc['''local'''], '''title''': doc['''title''']} ) else: new_doc_list.append(A__ ) __lowercase = new_doc_list __lowercase = [key for key, value in counts.items() if value > 1] __lowercase = [] for duplicate_key in duplicates: __lowercase = list({doc['''title'''] for doc in doc_list if doc['''local'''] == duplicate_key} ) if len(A__ ) > 1: raise ValueError( F"{duplicate_key} is present several times in the documentation table of content at " '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if '''local''' not in counts or counts[doc['''local''']] == 1] ) __lowercase = sorted(A__ , key=lambda A__ : s["title"].lower() ) # "overview" gets special treatment and is always first if len(A__ ) > 1: raise ValueError('''{doc_list} has two \'overview\' docs which is not allowed.''' ) overview_doc.extend(A__ ) # Sort return overview_doc def _A ( A__=False ): """simple docstring""" with open(A__ , encoding='''utf-8''' ) as f: __lowercase = yaml.safe_load(f.read() ) # Get to the API doc __lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 __lowercase = content[api_idx]['''sections'''] # Then to the model doc __lowercase = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 __lowercase = api_doc[scheduler_idx]['''sections'''] __lowercase = clean_doc_toc(A__ ) __lowercase = False if new_scheduler_doc != scheduler_doc: __lowercase = True if overwrite: __lowercase = new_scheduler_doc if diff: if overwrite: __lowercase = api_doc with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(A__ , allow_unicode=A__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) def _A ( A__=False ): """simple docstring""" with open(A__ , encoding='''utf-8''' ) as f: __lowercase = yaml.safe_load(f.read() ) # Get to the API doc __lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 __lowercase = content[api_idx]['''sections'''] # Then to the model doc __lowercase = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 __lowercase = False __lowercase = api_doc[pipeline_idx]['''sections'''] __lowercase = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: __lowercase = pipeline_doc['''section'''] __lowercase = clean_doc_toc(A__ ) if overwrite: __lowercase = new_sub_pipeline_doc new_pipeline_docs.append(A__ ) # sort overall pipeline doc __lowercase = clean_doc_toc(A__ ) if new_pipeline_docs != pipeline_docs: __lowercase = True if overwrite: __lowercase = new_pipeline_docs if diff: if overwrite: __lowercase = api_doc with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(A__ , allow_unicode=A__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') lowerCAmelCase__ = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Any ,lowercase__ : Any=1_3 ,lowercase__ : Dict=3_0 ,lowercase__ : Any=2 ,lowercase__ : Any=3 ,lowercase__ : Dict=True ,lowercase__ : int=True ,lowercase__ : Optional[int]=3_2 ,lowercase__ : List[Any]=2 ,lowercase__ : List[Any]=4 ,lowercase__ : Tuple=3_7 ,lowercase__ : Dict="gelu" ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Dict=0.1 ,lowercase__ : str=1_0 ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : int=3 ,lowercase__ : str=0.6 ,lowercase__ : Tuple=None ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = mask_ratio __lowercase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Dict ): return ViTMAEConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,decoder_hidden_size=self.hidden_size ,decoder_num_hidden_layers=self.num_hidden_layers ,decoder_num_attention_heads=self.num_attention_heads ,decoder_intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,mask_ratio=self.mask_ratio ,) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[int] ): __lowercase = TFViTMAEModel(config=lowercase__ ) __lowercase = model(lowercase__ ,training=lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : List[str] ): __lowercase = TFViTMAEForPreTraining(lowercase__ ) __lowercase = model(lowercase__ ,training=lowercase__ ) # expected sequence length = num_patches __lowercase = (self.image_size // self.patch_size) ** 2 __lowercase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) # test greyscale images __lowercase = 1 __lowercase = TFViTMAEForPreTraining(lowercase__ ) __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase__ ,training=lowercase__ ) __lowercase = self.patch_size**2 self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.prepare_config_and_inputs() ((__lowercase) , (__lowercase) , (__lowercase)) = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () SCREAMING_SNAKE_CASE : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = TFViTMAEModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(tf.keras.layers.Layer) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,tf.keras.layers.Layer ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # make the mask reproducible np.random.seed(2 ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = int((config.image_size // config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = model(lowercase__ ,noise=lowercase__ ) __lowercase = copy.deepcopy(self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = model(**lowercase__ ,noise=lowercase__ ) __lowercase = outputs_dict[0].numpy() __lowercase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) ,1e-6 ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # make the mask reproducible np.random.seed(2 ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = int((config.image_size // config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowercase__ : Tuple ): __lowercase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowercase__ ): __lowercase = v.numpy() else: __lowercase = np.array(lowercase__ ) return inputs_np_dict for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = prepare_numpy_arrays(lowercase__ ) __lowercase = model(lowercase__ ,noise=lowercase__ ) __lowercase = model(**lowercase__ ,noise=lowercase__ ) self.assert_outputs_same(lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Any ,lowercase__ : Tuple ,lowercase__ : List[str] ): # make masks reproducible np.random.seed(2 ) __lowercase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __lowercase = tf.constant(lowercase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument __lowercase = tf_noise super().check_pt_tf_models(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # make mask reproducible np.random.seed(2 ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowercase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowercase__ ,lowercase__ ),) if isinstance(lowercase__ ,lowercase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowercase__ ,'''_keras_serializable''' ,lowercase__ ) } __lowercase = int((config.image_size // config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __lowercase = tf.convert_to_tensor(lowercase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: __lowercase = main_layer_class(lowercase__ ) __lowercase = { name: tf.keras.Input(tensor.shape[1:] ,dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } __lowercase = tf.keras.Model(lowercase__ ,outputs=main_layer(lowercase__ ) ) __lowercase = model(lowercase__ ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''keras_model.h5''' ) model.save(lowercase__ ) __lowercase = tf.keras.models.load_model( lowercase__ ,custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowercase__ ,tf.keras.Model ) __lowercase = model(lowercase__ ) self.assert_outputs_same(lowercase__ ,lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ): # make mask reproducible np.random.seed(2 ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = int((config.image_size // config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = model(lowercase__ ,noise=lowercase__ ) if model_class.__name__ == "TFViTMAEModel": __lowercase = outputs.last_hidden_state.numpy() __lowercase = 0 else: __lowercase = outputs.logits.numpy() __lowercase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase__ ,saved_model=lowercase__ ) __lowercase = model_class.from_pretrained(lowercase__ ) __lowercase = model(lowercase__ ,noise=lowercase__ ) if model_class.__name__ == "TFViTMAEModel": __lowercase = after_outputs['''last_hidden_state'''].numpy() __lowercase = 0 else: __lowercase = after_outputs['''logits'''].numpy() __lowercase = 0 __lowercase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase__ ,1e-5 ) def SCREAMING_SNAKE_CASE ( self : Any ): # make mask reproducible np.random.seed(2 ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = int((config.image_size // config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = model(lowercase__ ,noise=lowercase__ ) __lowercase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowercase__ ) __lowercase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config __lowercase = model_class.from_config(model.config ) __lowercase = new_model(lowercase__ ) # Build model new_model.set_weights(model.get_weights() ) __lowercase = new_model(lowercase__ ,noise=lowercase__ ) self.assert_outputs_same(lowercase__ ,lowercase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : int ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE ( self : Dict ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) __lowercase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) __lowercase = ViTMAEConfig() __lowercase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(1, num_patches) ) # forward pass __lowercase = model(**lowercase__ ,noise=lowercase__ ) # verify the logits __lowercase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] ,lowercase__ ,atol=1e-4 )
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = inspect.getfile(accelerate.test_utils ) __lowercase = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_script.py'''] ) __lowercase = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_distributed_data_loop.py'''] ) __lowercase = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_ops.py'''] ) @require_multi_gpu def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): print(F"Found {torch.cuda.device_count()} devices." ) __lowercase = ['''torchrun''', F"--nproc_per_node={torch.cuda.device_count()}", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowercase__ ,env=os.environ.copy() ) @require_multi_gpu def SCREAMING_SNAKE_CASE ( self : int ): print(F"Found {torch.cuda.device_count()} devices." ) __lowercase = ['''torchrun''', F"--nproc_per_node={torch.cuda.device_count()}", self.operation_file_path] print(F"Command: {cmd}" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowercase__ ,env=os.environ.copy() ) @require_multi_gpu def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = ['''torchrun''', F"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowercase__ ,env=os.environ.copy() ) @require_multi_gpu def SCREAMING_SNAKE_CASE ( self : List[str] ): print(F"Found {torch.cuda.device_count()} devices, using 2 devices only" ) __lowercase = ['''torchrun''', F"--nproc_per_node={torch.cuda.device_count()}", self.data_loop_file_path] with patch_environment(omp_num_threads=1 ,cuda_visible_devices='''0,1''' ): execute_subprocess_async(lowercase__ ,env=os.environ.copy() ) if __name__ == "__main__": lowerCAmelCase__ = Accelerator() lowerCAmelCase__ = (accelerator.state.process_index + 2, 10) lowerCAmelCase__ = torch.randint(0, 10, shape).to(accelerator.device) lowerCAmelCase__ = '''''' lowerCAmelCase__ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." lowerCAmelCase__ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." lowerCAmelCase__ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def _A ( A__ , A__ , A__ , A__ , A__ = None , A__ = None , A__ = None , ): """simple docstring""" if config_name_or_path is None: __lowercase = '''facebook/rag-token-base''' if model_type == '''rag_token''' else '''facebook/rag-sequence-base''' if generator_tokenizer_name_or_path is None: __lowercase = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: __lowercase = question_encoder_name_or_path __lowercase = RagTokenForGeneration if model_type == '''rag_token''' else RagSequenceForGeneration # Save model. __lowercase = RagConfig.from_pretrained(A__ ) __lowercase = AutoConfig.from_pretrained(A__ ) __lowercase = AutoConfig.from_pretrained(A__ ) __lowercase = gen_config __lowercase = question_encoder_config __lowercase = model_class.from_pretrained_question_encoder_generator( A__ , A__ , config=A__ ) rag_model.save_pretrained(A__ ) # Sanity check. model_class.from_pretrained(A__ ) # Save tokenizers. __lowercase = AutoTokenizer.from_pretrained(A__ ) gen_tokenizer.save_pretrained(dest_dir / '''generator_tokenizer/''' ) __lowercase = AutoTokenizer.from_pretrained(A__ ) question_encoder_tokenizer.save_pretrained(dest_dir / '''question_encoder_tokenizer/''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--model_type''', choices=['''rag_sequence''', '''rag_token'''], required=True, type=str, help='''RAG model type: rag_sequence, rag_token''', ) parser.add_argument('''--dest''', type=str, required=True, help='''Path to the output checkpoint directory.''') parser.add_argument('''--generator_name_or_path''', type=str, required=True, help='''Generator model identifier''') parser.add_argument( '''--question_encoder_name_or_path''', type=str, required=True, help='''Question encoder model identifier''' ) parser.add_argument( '''--generator_tokenizer_name_or_path''', type=str, help='''Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``''', ) parser.add_argument( '''--question_encoder_tokenizer_name_or_path''', type=str, help='''Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``''', ) parser.add_argument( '''--config_name_or_path''', type=str, help=( '''Identifier of the model config to use, if not provided, resolves to a base config for a given''' ''' ``model_type``''' ), ) lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : int ,lowercase__ : Optional[int]=2 ,lowercase__ : Optional[Any]=3_2 ,lowercase__ : Optional[int]=1_6 ,lowercase__ : Optional[Any]=3 ,lowercase__ : List[Any]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Union[str, Any]=3_2 ,lowercase__ : Optional[int]=4 ,lowercase__ : List[str]=[0, 1, 2, 3] ,lowercase__ : List[str]=4 ,lowercase__ : List[str]=3_7 ,lowercase__ : List[Any]="gelu" ,lowercase__ : Optional[Any]=0.1 ,lowercase__ : Any=0.1 ,lowercase__ : int=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : Optional[int]=[1, 3_8_4, 2_4, 2_4] ,lowercase__ : Optional[int]=True ,lowercase__ : Tuple=None ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = backbone_out_indices __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = num_labels __lowercase = backbone_featmap_shape __lowercase = scope __lowercase = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 1 def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, '''hidden_sizes''': [9_6, 1_9_2, 3_8_4, 7_6_8], '''num_groups''': 2, } return DPTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,backbone_out_indices=self.backbone_out_indices ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,is_hybrid=self.is_hybrid ,backbone_config=lowercase__ ,backbone_featmap_shape=self.backbone_featmap_shape ,) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ): __lowercase = DPTModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : str ): __lowercase = self.num_labels __lowercase = DPTForDepthEstimation(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual(result.predicted_depth.shape ,(self.batch_size, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : int ): __lowercase = self.num_labels __lowercase = DPTForSemanticSegmentation(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE : Dict = ( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : str = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = DPTModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.config_tester.run_common_tests() @unittest.skip(reason='''DPT does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): pass def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True if model_class in get_values(lowercase__ ): continue __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : List[Any] ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = False __lowercase = True if model_class in get_values(lowercase__ ) or not model_class.supports_gradient_checkpointing: continue __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.gradient_checkpointing_enable() model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = _config_zero_init(lowercase__ ) for model_class in self.all_model_classes: __lowercase = model_class(config=lowercase__ ) # Skip the check for the backbone __lowercase = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": __lowercase = [F"{name}.{key}" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=F"Parameter {name} of model {model_class} seems not properly initialized" ,) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: __lowercase = DPTModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = '''add''' with self.assertRaises(lowercase__ ): __lowercase = DPTForDepthEstimation(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision @slow class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = DPTImageProcessor.from_pretrained('''Intel/dpt-hybrid-midas''' ) __lowercase = DPTForDepthEstimation.from_pretrained('''Intel/dpt-hybrid-midas''' ).to(lowercase__ ) __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) __lowercase = outputs.predicted_depth # verify the predicted depth __lowercase = torch.Size((1, 3_8_4, 3_8_4) ) self.assertEqual(predicted_depth.shape ,lowercase__ ) __lowercase = torch.tensor( [[[5.6_4_3_7, 5.6_1_4_6, 5.6_5_1_1], [5.4_3_7_1, 5.5_6_4_9, 5.5_9_5_8], [5.5_2_1_5, 5.5_1_8_4, 5.5_2_9_3]]] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 1_0_0 ,lowercase__ ,atol=1e-4 ) )
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def _A ( A__ = 8 ): __lowercase = ascii_letters + digits + punctuation return "".join(secrets.choice(A__ ) for _ in range(A__ ) ) def _A ( A__ , A__ ): i -= len(A__ ) __lowercase = i // 3 __lowercase = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) __lowercase = ( chars_incl + random(A__ , quotient + remainder ) + random(A__ , A__ ) + random(A__ , A__ ) ) __lowercase = list(A__ ) shuffle(A__ ) return "".join(A__ ) # random is a generalised function for letters, characters and numbers def _A ( A__ , A__ ): return "".join(secrets.choice(A__ ) for _ in range(A__ ) ) def _A ( A__ , A__ ): pass # Put your code here... def _A ( A__ , A__ ): pass # Put your code here... def _A ( A__ , A__ ): pass # Put your code here... def _A ( A__ , A__ = 8 ): if len(A__ ) < min_length: # Your Password must be at least 8 characters long return False __lowercase = any(char in ascii_uppercase for char in password ) __lowercase = any(char in ascii_lowercase for char in password ) __lowercase = any(char in digits for char in password ) __lowercase = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def _A ( ): __lowercase = int(input('''Please indicate the max length of your password: ''' ).strip() ) __lowercase = input( '''Please indicate the characters that must be in your password: ''' ).strip() print('''Password generated:''' , password_generator(A__ ) ) print( '''Alternative Password generated:''' , alternative_password_generator(A__ , A__ ) , ) print('''[If you are thinking of using this passsword, You better save it.]''' ) if __name__ == "__main__": main()
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def _A ( A__ ): """simple docstring""" return {key.lstrip('''-''' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def _A ( ): """simple docstring""" __lowercase = ArgumentParser( '''HuggingFace Datasets CLI tool''' , usage='''datasets-cli <command> [<args>]''' , allow_abbrev=A__ ) __lowercase = parser.add_subparsers(help='''datasets-cli command helpers''' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(A__ ) EnvironmentCommand.register_subcommand(A__ ) TestCommand.register_subcommand(A__ ) RunBeamCommand.register_subcommand(A__ ) DummyDataCommand.register_subcommand(A__ ) # Parse args __lowercase , __lowercase = parser.parse_known_args() if not hasattr(A__ , '''func''' ): parser.print_help() exit(1 ) __lowercase = parse_unknown_args(A__ ) # Run __lowercase = args.func(A__ , **A__ ) service.run() if __name__ == "__main__": main()
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' from ...processing_utils import ProcessorMixin class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = 'SpeechT5FeatureExtractor' SCREAMING_SNAKE_CASE : Tuple = 'SpeechT5Tokenizer' def __init__( self : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[Any] ): super().__init__(lowercase__ ,lowercase__ ) def __call__( self : List[Any] ,*lowercase__ : Optional[int] ,**lowercase__ : Union[str, Any] ): __lowercase = kwargs.pop('''audio''' ,lowercase__ ) __lowercase = kwargs.pop('''text''' ,lowercase__ ) __lowercase = kwargs.pop('''text_target''' ,lowercase__ ) __lowercase = kwargs.pop('''audio_target''' ,lowercase__ ) __lowercase = kwargs.pop('''sampling_rate''' ,lowercase__ ) if audio is not None and text is not None: raise ValueError( '''Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?''' ) if audio_target is not None and text_target is not None: raise ValueError( '''Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?''' ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( '''You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.''' ) if audio is not None: __lowercase = self.feature_extractor(lowercase__ ,*lowercase__ ,sampling_rate=lowercase__ ,**lowercase__ ) elif text is not None: __lowercase = self.tokenizer(lowercase__ ,**lowercase__ ) else: __lowercase = None if audio_target is not None: __lowercase = self.feature_extractor(audio_target=lowercase__ ,*lowercase__ ,sampling_rate=lowercase__ ,**lowercase__ ) __lowercase = targets['''input_values'''] elif text_target is not None: __lowercase = self.tokenizer(lowercase__ ,**lowercase__ ) __lowercase = targets['''input_ids'''] else: __lowercase = None if inputs is None: return targets if targets is not None: __lowercase = labels __lowercase = targets.get('''attention_mask''' ) if decoder_attention_mask is not None: __lowercase = decoder_attention_mask return inputs def SCREAMING_SNAKE_CASE ( self : str ,*lowercase__ : int ,**lowercase__ : Any ): __lowercase = kwargs.pop('''input_values''' ,lowercase__ ) __lowercase = kwargs.pop('''input_ids''' ,lowercase__ ) __lowercase = kwargs.pop('''labels''' ,lowercase__ ) if input_values is not None and input_ids is not None: raise ValueError('''Cannot process both `input_values` and `input_ids` inputs.''' ) if input_values is None and input_ids is None and labels is None: raise ValueError( '''You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.''' ) if input_values is not None: __lowercase = self.feature_extractor.pad(lowercase__ ,*lowercase__ ,**lowercase__ ) elif input_ids is not None: __lowercase = self.tokenizer.pad(lowercase__ ,**lowercase__ ) else: __lowercase = None if labels is not None: if "input_ids" in labels or (isinstance(lowercase__ ,lowercase__ ) and "input_ids" in labels[0]): __lowercase = self.tokenizer.pad(lowercase__ ,**lowercase__ ) __lowercase = targets['''input_ids'''] else: __lowercase = self.feature_extractor.feature_size __lowercase = self.feature_extractor.num_mel_bins __lowercase = self.feature_extractor.pad(lowercase__ ,*lowercase__ ,**lowercase__ ) __lowercase = feature_size_hack __lowercase = targets['''input_values'''] else: __lowercase = None if inputs is None: return targets if targets is not None: __lowercase = labels __lowercase = targets.get('''attention_mask''' ) if decoder_attention_mask is not None: __lowercase = decoder_attention_mask return inputs def SCREAMING_SNAKE_CASE ( self : List[Any] ,*lowercase__ : str ,**lowercase__ : str ): return self.tokenizer.batch_decode(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,*lowercase__ : Tuple ,**lowercase__ : List[Any] ): return self.tokenizer.decode(*lowercase__ ,**lowercase__ )
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" __lowercase = 1 for i in range(1 , num + 1 ): fact *= i return fact def _A ( A__ ): """simple docstring""" __lowercase = 0 while number > 0: __lowercase = number % 10 sum_of_digits += last_digit __lowercase = number // 10 # Removing the last_digit from the given number return sum_of_digits def _A ( A__ = 100 ): """simple docstring""" __lowercase = factorial(A__ ) __lowercase = split_and_add(A__ ) return result if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[int] ,lowercase__ : WhisperForConditionalGeneration ,lowercase__ : WhisperProcessor ,lowercase__ : AutoencoderKL ,lowercase__ : CLIPTextModel ,lowercase__ : CLIPTokenizer ,lowercase__ : UNetaDConditionModel ,lowercase__ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] ,lowercase__ : StableDiffusionSafetyChecker ,lowercase__ : CLIPImageProcessor ,): super().__init__() if safety_checker is None: logger.warning( F"You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure" ''' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered''' ''' results in services or applications open to the public. Both the diffusers team and Hugging Face''' ''' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling''' ''' it only for use-cases that involve analyzing network behavior or auditing its results. For more''' ''' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .''' ) self.register_modules( speech_model=lowercase__ ,speech_processor=lowercase__ ,vae=lowercase__ ,text_encoder=lowercase__ ,tokenizer=lowercase__ ,unet=lowercase__ ,scheduler=lowercase__ ,feature_extractor=lowercase__ ,) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Union[str, int]] = "auto" ): if slice_size == "auto": __lowercase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): self.enable_attention_slicing(lowercase__ ) @torch.no_grad() def __call__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : int=1_6_0_0_0 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Union[str, Any] ,): __lowercase = self.speech_processor.feature_extractor( lowercase__ ,return_tensors='''pt''' ,sampling_rate=lowercase__ ).input_features.to(self.device ) __lowercase = self.speech_model.generate(lowercase__ ,max_length=4_8_0_0_0_0 ) __lowercase = self.speech_processor.tokenizer.batch_decode(lowercase__ ,skip_special_tokens=lowercase__ ,normalize=lowercase__ )[ 0 ] if isinstance(lowercase__ ,lowercase__ ): __lowercase = 1 elif isinstance(lowercase__ ,lowercase__ ): __lowercase = len(lowercase__ ) else: raise ValueError(F"`prompt` has to be of type `str` or `list` but is {type(lowercase__ )}" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` have to be divisible by 8 but are {height} and {width}." ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(lowercase__ ,lowercase__ ) or callback_steps <= 0) ): raise ValueError( F"`callback_steps` has to be a positive integer but is {callback_steps} of type" F" {type(lowercase__ )}." ) # get prompt text embeddings __lowercase = self.tokenizer( lowercase__ ,padding='''max_length''' ,max_length=self.tokenizer.model_max_length ,return_tensors='''pt''' ,) __lowercase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: __lowercase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F" {self.tokenizer.model_max_length} tokens: {removed_text}" ) __lowercase = text_input_ids[:, : self.tokenizer.model_max_length] __lowercase = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method __lowercase , __lowercase , __lowercase = text_embeddings.shape __lowercase = text_embeddings.repeat(1 ,lowercase__ ,1 ) __lowercase = text_embeddings.view(bs_embed * num_images_per_prompt ,lowercase__ ,-1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. __lowercase = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: __lowercase = 4_2 if negative_prompt is None: __lowercase = [''''''] * batch_size elif type(lowercase__ ) is not type(lowercase__ ): raise TypeError( F"`negative_prompt` should be the same type to `prompt`, but got {type(lowercase__ )} !=" F" {type(lowercase__ )}." ) elif isinstance(lowercase__ ,lowercase__ ): __lowercase = [negative_prompt] elif batch_size != len(lowercase__ ): raise ValueError( F"`negative_prompt`: {negative_prompt} has batch size {len(lowercase__ )}, but `prompt`:" F" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches" ''' the batch size of `prompt`.''' ) else: __lowercase = negative_prompt __lowercase = text_input_ids.shape[-1] __lowercase = self.tokenizer( lowercase__ ,padding='''max_length''' ,max_length=lowercase__ ,truncation=lowercase__ ,return_tensors='''pt''' ,) __lowercase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __lowercase = uncond_embeddings.shape[1] __lowercase = uncond_embeddings.repeat(1 ,lowercase__ ,1 ) __lowercase = uncond_embeddings.view(batch_size * num_images_per_prompt ,lowercase__ ,-1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __lowercase = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. __lowercase = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) __lowercase = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps __lowercase = torch.randn(lowercase__ ,generator=lowercase__ ,device='''cpu''' ,dtype=lowercase__ ).to( self.device ) else: __lowercase = torch.randn(lowercase__ ,generator=lowercase__ ,device=self.device ,dtype=lowercase__ ) else: if latents.shape != latents_shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) __lowercase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(lowercase__ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand __lowercase = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __lowercase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __lowercase = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __lowercase = {} if accepts_eta: __lowercase = eta for i, t in enumerate(self.progress_bar(lowercase__ ) ): # expand the latents if we are doing classifier free guidance __lowercase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __lowercase = self.scheduler.scale_model_input(lowercase__ ,lowercase__ ) # predict the noise residual __lowercase = self.unet(lowercase__ ,lowercase__ ,encoder_hidden_states=lowercase__ ).sample # perform guidance if do_classifier_free_guidance: __lowercase , __lowercase = noise_pred.chunk(2 ) __lowercase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 __lowercase = self.scheduler.step(lowercase__ ,lowercase__ ,lowercase__ ,**lowercase__ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(lowercase__ ,lowercase__ ,lowercase__ ) __lowercase = 1 / 0.1_8_2_1_5 * latents __lowercase = self.vae.decode(lowercase__ ).sample __lowercase = (image / 2 + 0.5).clamp(0 ,1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 __lowercase = image.cpu().permute(0 ,2 ,3 ,1 ).float().numpy() if output_type == "pil": __lowercase = self.numpy_to_pil(lowercase__ ) if not return_dict: return image return StableDiffusionPipelineOutput(images=lowercase__ ,nsfw_content_detected=lowercase__ )
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from __future__ import annotations def _A ( A__ ): """simple docstring""" return [ord(A__ ) - 96 for elem in plain] def _A ( A__ ): """simple docstring""" return "".join(chr(elem + 96 ) for elem in encoded ) def _A ( ): """simple docstring""" __lowercase = encode(input('''-> ''' ).strip().lower() ) print('''Encoded: ''' , A__ ) print('''Decoded:''' , decode(A__ ) ) if __name__ == "__main__": main()
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' from __future__ import annotations lowerCAmelCase__ = 1.6_021e-19 # units = C def _A ( A__ , A__ , A__ , ): """simple docstring""" if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = CustomTokenizer pass
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' import gc import threading import time import psutil import torch class lowercase_ : """simple docstring""" def __init__( self : List[Any] ): __lowercase = psutil.Process() __lowercase = False def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = -1 while True: __lowercase = max(self.process.memory_info().rss ,self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = True __lowercase = threading.Thread(target=self.peak_monitor ) __lowercase = True self.thread.start() def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = False self.thread.join() return self.cpu_memory_peak lowerCAmelCase__ = PeakCPUMemory() def _A ( ): """simple docstring""" __lowercase = {'''time''': time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem __lowercase = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): __lowercase = torch.cuda.memory_allocated(A__ ) torch.cuda.reset_peak_memory_stats() return measures def _A ( A__ ): """simple docstring""" __lowercase = {'''time''': time.time() - start_measures['''time''']} gc.collect() torch.cuda.empty_cache() # CPU mem __lowercase = (psutil.Process().memory_info().rss - start_measures['''cpu''']) / 2**20 __lowercase = (cpu_peak_tracker.stop() - start_measures['''cpu''']) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): __lowercase = (torch.cuda.memory_allocated(A__ ) - start_measures[str(A__ )]) / 2**20 __lowercase = (torch.cuda.max_memory_allocated(A__ ) - start_measures[str(A__ )]) / 2**20 return measures def _A ( A__ , A__ ): """simple docstring""" print(F"{description}:" ) print(F"- Time: {measures['time']:.2f}s" ) for i in range(torch.cuda.device_count() ): print(F"- GPU {i} allocated: {measures[str(A__ )]:.2f}MiB" ) __lowercase = measures[F"{i}-peak"] print(F"- GPU {i} peak: {peak:.2f}MiB" ) print(F"- CPU RAM allocated: {measures['cpu']:.2f}MiB" ) print(F"- CPU RAM peak: {measures['cpu-peak']:.2f}MiB" )
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
import warnings from ..trainer import Trainer from ..utils import logging lowerCAmelCase__ = logging.get_logger(__name__) class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Any ,lowercase__ : Any=None ,**lowercase__ : Optional[int] ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' ,lowercase__ ,) super().__init__(args=lowercase__ ,**lowercase__ )
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowercase_ (unittest.TestCase ): def __init__( self : Optional[int] ,lowercase__ : List[str] ,lowercase__ : List[Any]=7 ,lowercase__ : int=3 ,lowercase__ : Dict=1_8 ,lowercase__ : Optional[int]=3_0 ,lowercase__ : str=4_0_0 ,lowercase__ : Optional[Any]=True ,lowercase__ : Any=None ,lowercase__ : Any=True ,lowercase__ : Tuple=None ,): __lowercase = size if size is not None else {'''shortest_edge''': 2_0} __lowercase = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = do_center_crop __lowercase = crop_size def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class lowercase_ (lowerCamelCase__ , unittest.TestCase ): SCREAMING_SNAKE_CASE : Tuple = MobileNetVaImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = MobileNetVaImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : str ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase__ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''size''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_center_crop''' ) ) self.assertTrue(hasattr(lowercase__ ,'''crop_size''' ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size ,{'''height''': 1_8, '''width''': 1_8} ) __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ,crop_size=8_4 ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size ,{'''height''': 8_4, '''width''': 8_4} ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,numpify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Tuple ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,torchify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,)
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' import math def _A ( A__ ): """simple docstring""" if not isinstance(A__ , A__ ): __lowercase = F"Input value of [number={number}] must be an integer" raise TypeError(A__ ) if number < 1: __lowercase = F"Input value of [number={number}] must be > 0" raise ValueError(A__ ) elif number == 1: return 3 elif number == 2: return 5 else: __lowercase = int(math.log(number // 3 , 2 ) ) + 2 __lowercase = [3, 5] __lowercase = 2 __lowercase = 3 for block in range(1 , A__ ): for _ in range(A__ ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): lowerCAmelCase__ = 0 try: lowerCAmelCase__ = proth(number) except ValueError: print(f'ValueError: there is no {number}th Proth number') continue print(f'The {number}th Proth number: {value}')
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' import math def _A ( A__ ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(A__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _A ( A__ = 10001 ): """simple docstring""" try: __lowercase = int(A__ ) except (TypeError, ValueError): raise TypeError('''Parameter nth must be int or castable to int.''' ) from None if nth <= 0: raise ValueError('''Parameter nth must be greater than or equal to one.''' ) __lowercase = [] __lowercase = 2 while len(A__ ) < nth: if is_prime(A__ ): primes.append(A__ ) num += 1 else: num += 1 return primes[len(A__ ) - 1] if __name__ == "__main__": print(f'{solution() = }')
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : NestedDataStructureLike[PathLike] ,lowercase__ : Optional[NamedSplit] = None ,lowercase__ : Optional[Features] = None ,lowercase__ : str = None ,lowercase__ : bool = False ,lowercase__ : bool = False ,lowercase__ : Optional[str] = None ,lowercase__ : Optional[int] = None ,**lowercase__ : Optional[int] ,): super().__init__( lowercase__ ,split=lowercase__ ,features=lowercase__ ,cache_dir=lowercase__ ,keep_in_memory=lowercase__ ,streaming=lowercase__ ,num_proc=lowercase__ ,**lowercase__ ,) __lowercase = field __lowercase = path_or_paths if isinstance(lowercase__ ,lowercase__ ) else {self.split: path_or_paths} __lowercase = Json( cache_dir=lowercase__ ,data_files=lowercase__ ,features=lowercase__ ,field=lowercase__ ,**lowercase__ ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Build iterable dataset if self.streaming: __lowercase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __lowercase = None __lowercase = None __lowercase = None __lowercase = None self.builder.download_and_prepare( download_config=lowercase__ ,download_mode=lowercase__ ,verification_mode=lowercase__ ,base_path=lowercase__ ,num_proc=self.num_proc ,) __lowercase = self.builder.as_dataset( split=self.split ,verification_mode=lowercase__ ,in_memory=self.keep_in_memory ) return dataset class lowercase_ : """simple docstring""" def __init__( self : str ,lowercase__ : Dataset ,lowercase__ : Union[PathLike, BinaryIO] ,lowercase__ : Optional[int] = None ,lowercase__ : Optional[int] = None ,**lowercase__ : Any ,): if num_proc is not None and num_proc <= 0: raise ValueError(F"num_proc {num_proc} must be an integer > 0." ) __lowercase = dataset __lowercase = path_or_buf __lowercase = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __lowercase = num_proc __lowercase = '''utf-8''' __lowercase = to_json_kwargs def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.to_json_kwargs.pop('''path_or_buf''' ,lowercase__ ) __lowercase = self.to_json_kwargs.pop('''orient''' ,'''records''' ) __lowercase = self.to_json_kwargs.pop('''lines''' ,True if orient == '''records''' else False ) __lowercase = self.to_json_kwargs.pop('''index''' ,False if orient in ['''split''', '''table'''] else True ) __lowercase = self.to_json_kwargs.pop('''compression''' ,lowercase__ ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F"`datasets` currently does not support {compression} compression" ) if isinstance(self.path_or_buf ,(str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf ,'''wb''' ,compression=lowercase__ ) as buffer: __lowercase = self._write(file_obj=lowercase__ ,orient=lowercase__ ,lines=lowercase__ ,index=lowercase__ ,**self.to_json_kwargs ) else: if compression: raise NotImplementedError( F"The compression parameter is not supported when writing to a buffer, but compression={compression}" ''' was passed. Please provide a local path instead.''' ) __lowercase = self._write( file_obj=self.path_or_buf ,orient=lowercase__ ,lines=lowercase__ ,index=lowercase__ ,**self.to_json_kwargs ) return written def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : str ): __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = args __lowercase = query_table( table=self.dataset.data ,key=slice(lowercase__ ,offset + self.batch_size ) ,indices=self.dataset._indices ,) __lowercase = batch.to_pandas().to_json( path_or_buf=lowercase__ ,orient=lowercase__ ,lines=lowercase__ ,index=lowercase__ ,**lowercase__ ) if not json_str.endswith('''\n''' ): json_str += "\n" return json_str.encode(self.encoding ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : BinaryIO ,lowercase__ : int ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,**lowercase__ : Any ,): __lowercase = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 ,len(self.dataset ) ,self.batch_size ) ,unit='''ba''' ,disable=not logging.is_progress_bar_enabled() ,desc='''Creating json from Arrow format''' ,): __lowercase = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(lowercase__ ) else: __lowercase , __lowercase = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json ,[(offset, orient, lines, index, to_json_kwargs) for offset in range(0 ,lowercase__ ,lowercase__ )] ,) ,total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size ,unit='''ba''' ,disable=not logging.is_progress_bar_enabled() ,desc='''Creating json from Arrow format''' ,): written += file_obj.write(lowercase__ ) return written
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = '''ylacombe/bark-small''' __lowercase = tempfile.mkdtemp() __lowercase = '''en_speaker_1''' __lowercase = '''This is a test string''' __lowercase = '''speaker_embeddings_path.json''' __lowercase = '''speaker_embeddings''' def SCREAMING_SNAKE_CASE ( self : int ,**lowercase__ : Union[str, Any] ): return AutoTokenizer.from_pretrained(self.checkpoint ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.get_tokenizer() __lowercase = BarkProcessor(tokenizer=lowercase__ ) processor.save_pretrained(self.tmpdirname ) __lowercase = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint ,speaker_embeddings_dict_path=self.speaker_embeddings_dict_path ,) processor.save_pretrained( self.tmpdirname ,speaker_embeddings_dict_path=self.speaker_embeddings_dict_path ,speaker_embeddings_directory=self.speaker_embeddings_directory ,) __lowercase = self.get_tokenizer(bos_token='''(BOS)''' ,eos_token='''(EOS)''' ) __lowercase = BarkProcessor.from_pretrained( self.tmpdirname ,self.speaker_embeddings_dict_path ,bos_token='''(BOS)''' ,eos_token='''(EOS)''' ,) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint ,speaker_embeddings_dict_path=self.speaker_embeddings_dict_path ,) __lowercase = 3_5 __lowercase = 2 __lowercase = 8 __lowercase = { '''semantic_prompt''': np.ones(lowercase__ ), '''coarse_prompt''': np.ones((nb_codebooks_coarse, seq_len) ), '''fine_prompt''': np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset __lowercase = processor(text=self.input_string ,voice_preset=lowercase__ ) __lowercase = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() ,processed_voice_preset.get(lowercase__ ,np.array([] ) ).tolist() ) # test loading voice preset from npz file __lowercase = os.path.join(self.tmpdirname ,'''file.npz''' ) np.savez(lowercase__ ,**lowercase__ ) __lowercase = processor(text=self.input_string ,voice_preset=lowercase__ ) __lowercase = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() ,processed_voice_preset.get(lowercase__ ,np.array([] ) ).tolist() ) # test loading voice preset from the hub __lowercase = processor(text=self.input_string ,voice_preset=self.voice_preset ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.get_tokenizer() __lowercase = BarkProcessor(tokenizer=lowercase__ ) __lowercase = processor(text=self.input_string ) __lowercase = tokenizer( self.input_string ,padding='''max_length''' ,max_length=2_5_6 ,add_special_tokens=lowercase__ ,return_attention_mask=lowercase__ ,return_token_type_ids=lowercase__ ,) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key].squeeze().tolist() )
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# lowerCAmelCase__ = [ # (stable-diffusion, HF Diffusers) ('''time_embed.0.weight''', '''time_embedding.linear_1.weight'''), ('''time_embed.0.bias''', '''time_embedding.linear_1.bias'''), ('''time_embed.2.weight''', '''time_embedding.linear_2.weight'''), ('''time_embed.2.bias''', '''time_embedding.linear_2.bias'''), ('''input_blocks.0.0.weight''', '''conv_in.weight'''), ('''input_blocks.0.0.bias''', '''conv_in.bias'''), ('''out.0.weight''', '''conv_norm_out.weight'''), ('''out.0.bias''', '''conv_norm_out.bias'''), ('''out.2.weight''', '''conv_out.weight'''), ('''out.2.bias''', '''conv_out.bias'''), ] lowerCAmelCase__ = [ # (stable-diffusion, HF Diffusers) ('''in_layers.0''', '''norm1'''), ('''in_layers.2''', '''conv1'''), ('''out_layers.0''', '''norm2'''), ('''out_layers.3''', '''conv2'''), ('''emb_layers.1''', '''time_emb_proj'''), ('''skip_connection''', '''conv_shortcut'''), ] lowerCAmelCase__ = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks lowerCAmelCase__ = f'down_blocks.{i}.resnets.{j}.' lowerCAmelCase__ = f'input_blocks.{3*i + j + 1}.0.' unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 lowerCAmelCase__ = f'down_blocks.{i}.attentions.{j}.' lowerCAmelCase__ = f'input_blocks.{3*i + j + 1}.1.' unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks lowerCAmelCase__ = f'up_blocks.{i}.resnets.{j}.' lowerCAmelCase__ = f'output_blocks.{3*i + j}.0.' unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 lowerCAmelCase__ = f'up_blocks.{i}.attentions.{j}.' lowerCAmelCase__ = f'output_blocks.{3*i + j}.1.' unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 lowerCAmelCase__ = f'down_blocks.{i}.downsamplers.0.conv.' lowerCAmelCase__ = f'input_blocks.{3*(i+1)}.0.op.' unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 lowerCAmelCase__ = f'up_blocks.{i}.upsamplers.0.' lowerCAmelCase__ = f'output_blocks.{3*i + 2}.{1 if i == 0 else 2}.' unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) lowerCAmelCase__ = '''mid_block.attentions.0.''' lowerCAmelCase__ = '''middle_block.1.''' unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): lowerCAmelCase__ = f'mid_block.resnets.{j}.' lowerCAmelCase__ = f'middle_block.{2*j}.' unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def _A ( A__ ): """simple docstring""" __lowercase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: __lowercase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: __lowercase = v.replace(A__ , A__ ) __lowercase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: __lowercase = v.replace(A__ , A__ ) __lowercase = v __lowercase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# lowerCAmelCase__ = [ # (stable-diffusion, HF Diffusers) ('''nin_shortcut''', '''conv_shortcut'''), ('''norm_out''', '''conv_norm_out'''), ('''mid.attn_1.''', '''mid_block.attentions.0.'''), ] for i in range(4): # down_blocks have two resnets for j in range(2): lowerCAmelCase__ = f'encoder.down_blocks.{i}.resnets.{j}.' lowerCAmelCase__ = f'encoder.down.{i}.block.{j}.' vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: lowerCAmelCase__ = f'down_blocks.{i}.downsamplers.0.' lowerCAmelCase__ = f'down.{i}.downsample.' vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) lowerCAmelCase__ = f'up_blocks.{i}.upsamplers.0.' lowerCAmelCase__ = f'up.{3-i}.upsample.' vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): lowerCAmelCase__ = f'decoder.up_blocks.{i}.resnets.{j}.' lowerCAmelCase__ = f'decoder.up.{3-i}.block.{j}.' vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): lowerCAmelCase__ = f'mid_block.resnets.{i}.' lowerCAmelCase__ = f'mid.block_{i+1}.' vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) lowerCAmelCase__ = [ # (stable-diffusion, HF Diffusers) ('''norm.''', '''group_norm.'''), ('''q.''', '''query.'''), ('''k.''', '''key.'''), ('''v.''', '''value.'''), ('''proj_out.''', '''proj_attn.'''), ] def _A ( A__ ): """simple docstring""" return w.reshape(*w.shape , 1 , 1 ) def _A ( A__ ): """simple docstring""" __lowercase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: __lowercase = v.replace(A__ , A__ ) __lowercase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: __lowercase = v.replace(A__ , A__ ) __lowercase = v __lowercase = {v: vae_state_dict[k] for k, v in mapping.items()} __lowercase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F"mid.attn_1.{weight_name}.weight" in k: print(F"Reshaping {k} for SD format" ) __lowercase = reshape_weight_for_sd(A__ ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# lowerCAmelCase__ = [ # (stable-diffusion, HF Diffusers) ('''resblocks.''', '''text_model.encoder.layers.'''), ('''ln_1''', '''layer_norm1'''), ('''ln_2''', '''layer_norm2'''), ('''.c_fc.''', '''.fc1.'''), ('''.c_proj.''', '''.fc2.'''), ('''.attn''', '''.self_attn'''), ('''ln_final.''', '''transformer.text_model.final_layer_norm.'''), ('''token_embedding.weight''', '''transformer.text_model.embeddings.token_embedding.weight'''), ('''positional_embedding''', '''transformer.text_model.embeddings.position_embedding.weight'''), ] lowerCAmelCase__ = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} lowerCAmelCase__ = re.compile('''|'''.join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp lowerCAmelCase__ = {'''q''': 0, '''k''': 1, '''v''': 2} def _A ( A__ ): """simple docstring""" __lowercase = {} __lowercase = {} __lowercase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): __lowercase = k[: -len('''.q_proj.weight''' )] __lowercase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: __lowercase = [None, None, None] __lowercase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): __lowercase = k[: -len('''.q_proj.bias''' )] __lowercase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: __lowercase = [None, None, None] __lowercase = v continue __lowercase = textenc_pattern.sub(lambda A__ : protected[re.escape(m.group(0 ) )] , A__ ) __lowercase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) __lowercase = textenc_pattern.sub(lambda A__ : protected[re.escape(m.group(0 ) )] , A__ ) __lowercase = torch.cat(A__ ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) __lowercase = textenc_pattern.sub(lambda A__ : protected[re.escape(m.group(0 ) )] , A__ ) __lowercase = torch.cat(A__ ) return new_state_dict def _A ( A__ ): """simple docstring""" return text_enc_dict if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--model_path''', default=None, type=str, required=True, help='''Path to the model to convert.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument('''--half''', action='''store_true''', help='''Save weights in half precision.''') parser.add_argument( '''--use_safetensors''', action='''store_true''', help='''Save weights use safetensors, default is ckpt.''' ) lowerCAmelCase__ = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors lowerCAmelCase__ = osp.join(args.model_path, '''unet''', '''diffusion_pytorch_model.safetensors''') lowerCAmelCase__ = osp.join(args.model_path, '''vae''', '''diffusion_pytorch_model.safetensors''') lowerCAmelCase__ = osp.join(args.model_path, '''text_encoder''', '''model.safetensors''') # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): lowerCAmelCase__ = load_file(unet_path, device='''cpu''') else: lowerCAmelCase__ = osp.join(args.model_path, '''unet''', '''diffusion_pytorch_model.bin''') lowerCAmelCase__ = torch.load(unet_path, map_location='''cpu''') if osp.exists(vae_path): lowerCAmelCase__ = load_file(vae_path, device='''cpu''') else: lowerCAmelCase__ = osp.join(args.model_path, '''vae''', '''diffusion_pytorch_model.bin''') lowerCAmelCase__ = torch.load(vae_path, map_location='''cpu''') if osp.exists(text_enc_path): lowerCAmelCase__ = load_file(text_enc_path, device='''cpu''') else: lowerCAmelCase__ = osp.join(args.model_path, '''text_encoder''', '''pytorch_model.bin''') lowerCAmelCase__ = torch.load(text_enc_path, map_location='''cpu''') # Convert the UNet model lowerCAmelCase__ = convert_unet_state_dict(unet_state_dict) lowerCAmelCase__ = {'''model.diffusion_model.''' + k: v for k, v in unet_state_dict.items()} # Convert the VAE model lowerCAmelCase__ = convert_vae_state_dict(vae_state_dict) lowerCAmelCase__ = {'''first_stage_model.''' + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper lowerCAmelCase__ = '''text_model.encoder.layers.22.layer_norm2.bias''' in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm lowerCAmelCase__ = {'''transformer.''' + k: v for k, v in text_enc_dict.items()} lowerCAmelCase__ = convert_text_enc_state_dict_vaa(text_enc_dict) lowerCAmelCase__ = {'''cond_stage_model.model.''' + k: v for k, v in text_enc_dict.items()} else: lowerCAmelCase__ = convert_text_enc_state_dict(text_enc_dict) lowerCAmelCase__ = {'''cond_stage_model.transformer.''' + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint lowerCAmelCase__ = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: lowerCAmelCase__ = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: lowerCAmelCase__ = {'''state_dict''': state_dict} torch.save(state_dict, args.checkpoint_path)
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = 'philschmid/bart-large-cnn-samsum' SCREAMING_SNAKE_CASE : str = ( 'This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ' 'and returns a summary of the text.' ) SCREAMING_SNAKE_CASE : List[str] = 'summarizer' SCREAMING_SNAKE_CASE : List[Any] = AutoTokenizer SCREAMING_SNAKE_CASE : str = AutoModelForSeqaSeqLM SCREAMING_SNAKE_CASE : List[str] = ['text'] SCREAMING_SNAKE_CASE : List[str] = ['text'] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ): return self.pre_processor(lowercase__ ,return_tensors='''pt''' ,truncation=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ): return self.model.generate(**lowercase__ )[0] def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Union[str, Any] ): return self.pre_processor.decode(lowercase__ ,skip_special_tokens=lowercase__ ,clean_up_tokenization_spaces=lowercase__ )
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowerCAmelCase__ = '''<<<<<<< This should probably be modified because it mentions: ''' lowerCAmelCase__ = '''======= >>>>>>> ''' lowerCAmelCase__ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] lowerCAmelCase__ = [ # (pattern, replacement) # Order is important here for some replacements (R'''tfds\.core''', R'''datasets'''), (R'''tf\.io\.gfile\.GFile''', R'''open'''), (R'''tf\.([\w\d]+)''', R'''datasets.Value(\'\1\')'''), (R'''tfds\.features\.Text\(\)''', R'''datasets.Value(\'string\')'''), (R'''tfds\.features\.Text\(''', R'''datasets.Value(\'string\'),'''), (R'''features\s*=\s*tfds.features.FeaturesDict\(''', R'''features=datasets.Features('''), (R'''tfds\.features\.FeaturesDict\(''', R'''dict('''), (R'''The TensorFlow Datasets Authors''', R'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (R'''tfds\.''', R'''datasets.'''), (R'''dl_manager\.manual_dir''', R'''self.config.data_dir'''), (R'''self\.builder_config''', R'''self.config'''), ] def _A ( A__ ): """simple docstring""" return ConvertCommand(args.tfds_path , args.datasets_directory ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser( '''convert''' ,help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' ,) train_parser.add_argument( '''--tfds_path''' ,type=lowercase__ ,required=lowercase__ ,help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' ,) train_parser.add_argument( '''--datasets_directory''' ,type=lowercase__ ,required=lowercase__ ,help='''Path to the HuggingFace Datasets folder.''' ) train_parser.set_defaults(func=lowercase__ ) def __init__( self : Optional[Any] ,lowercase__ : str ,lowercase__ : str ,*lowercase__ : Optional[Any] ): __lowercase = get_logger('''datasets-cli/converting''' ) __lowercase = tfds_path __lowercase = datasets_directory def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): if os.path.isdir(self._tfds_path ): __lowercase = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): __lowercase = os.path.dirname(self._tfds_path ) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''' ) __lowercase = os.path.abspath(self._datasets_directory ) self._logger.info(F"Converting datasets from {abs_tfds_path} to {abs_datasets_path}" ) __lowercase = [] __lowercase = [] __lowercase = {} if os.path.isdir(self._tfds_path ): __lowercase = os.listdir(lowercase__ ) else: __lowercase = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F"Looking at file {f_name}" ) __lowercase = os.path.join(lowercase__ ,lowercase__ ) __lowercase = os.path.join(lowercase__ ,lowercase__ ) if not os.path.isfile(lowercase__ ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''' ) continue with open(lowercase__ ,encoding='''utf-8''' ) as f: __lowercase = f.readlines() __lowercase = [] __lowercase = False __lowercase = False __lowercase = [] for line in lines: __lowercase = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: __lowercase = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here __lowercase = '''''' continue elif "from absl import logging" in out_line: __lowercase = '''from datasets import logging\n''' elif "getLogger" in out_line: __lowercase = out_line.replace('''getLogger''' ,'''get_logger''' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): __lowercase = True __lowercase = list(filter(lambda lowercase__ : e in out_line ,lowercase__ ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(lowercase__ ) + '''\n''' ) out_lines.append(lowercase__ ) out_lines.append(lowercase__ ) continue else: for pattern, replacement in TO_CONVERT: __lowercase = re.sub(lowercase__ ,lowercase__ ,lowercase__ ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: __lowercase = re.match(r'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' ,lowercase__ ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(''',''' ) ) __lowercase = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F"Error converting {out_line.strip()}" ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: __lowercase = True out_lines.append(lowercase__ ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset __lowercase = f_name.replace('''.py''' ,'''''' ) __lowercase = os.path.join(lowercase__ ,lowercase__ ) __lowercase = os.path.join(lowercase__ ,lowercase__ ) os.makedirs(lowercase__ ,exist_ok=lowercase__ ) self._logger.info(F"Adding directory {output_dir}" ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(lowercase__ ) if needs_manual_update: with_manual_update.append(lowercase__ ) with open(lowercase__ ,'''w''' ,encoding='''utf-8''' ) as f: f.writelines(lowercase__ ) self._logger.info(F"Converted in {output_file}" ) for utils_file in utils_files: try: __lowercase = os.path.basename(lowercase__ ) __lowercase = imports_to_builder_map[f_name.replace('''.py''' ,'''''' )] self._logger.info(F"Moving {dest_folder} to {utils_file}" ) shutil.copy(lowercase__ ,lowercase__ ) except KeyError: self._logger.error(F"Cannot find destination folder for {utils_file}. Please copy manually." ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F"You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'." )
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import math def _A ( A__ , A__ ): """simple docstring""" if ( not isinstance(A__ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * power_factor def _A ( A__ , A__ ): """simple docstring""" if ( not isinstance(A__ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def _A ( A__ ): """simple docstring""" if ( (cp >= 0x4e00 and cp <= 0x9fff) or (cp >= 0x3400 and cp <= 0x4dbf) # or (cp >= 0x2_0000 and cp <= 0x2_a6df) # or (cp >= 0x2_a700 and cp <= 0x2_b73f) # or (cp >= 0x2_b740 and cp <= 0x2_b81f) # or (cp >= 0x2_b820 and cp <= 0x2_ceaf) # or (cp >= 0xf900 and cp <= 0xfaff) or (cp >= 0x2_f800 and cp <= 0x2_fa1f) # ): # return True return False def _A ( A__ ): """simple docstring""" for char in word: __lowercase = ord(A__ ) if not _is_chinese_char(A__ ): return 0 return 1 def _A ( A__ ): """simple docstring""" __lowercase = set() for token in tokens: __lowercase = len(A__ ) > 1 and is_chinese(A__ ) if chinese_word: word_set.add(A__ ) __lowercase = list(A__ ) return word_list def _A ( A__ , A__ ): """simple docstring""" if not chinese_word_set: return bert_tokens __lowercase = max([len(A__ ) for w in chinese_word_set] ) __lowercase = bert_tokens __lowercase , __lowercase = 0, len(A__ ) while start < end: __lowercase = True if is_chinese(bert_word[start] ): __lowercase = min(end - start , A__ ) for i in range(A__ , 1 , -1 ): __lowercase = ''''''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): __lowercase = '''##''' + bert_word[j] __lowercase = start + i __lowercase = False break if single_word: start += 1 return bert_word def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = [] for i in range(0 , len(A__ ) , 100 ): __lowercase = ltp_tokenizer.seg(lines[i : i + 100] )[0] __lowercase = [get_chinese_word(A__ ) for r in res] ltp_res.extend(A__ ) assert len(A__ ) == len(A__ ) __lowercase = [] for i in range(0 , len(A__ ) , 100 ): __lowercase = bert_tokenizer(lines[i : i + 100] , add_special_tokens=A__ , truncation=A__ , max_length=512 ) bert_res.extend(res['''input_ids'''] ) assert len(A__ ) == len(A__ ) __lowercase = [] for input_ids, chinese_word in zip(A__ , A__ ): __lowercase = [] for id in input_ids: __lowercase = bert_tokenizer._convert_id_to_token(A__ ) input_tokens.append(A__ ) __lowercase = add_sub_symbol(A__ , A__ ) __lowercase = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(A__ ): if token[:2] == "##": __lowercase = token[2:] # save chinese tokens' pos if len(A__ ) == 1 and _is_chinese_char(ord(A__ ) ): ref_id.append(A__ ) ref_ids.append(A__ ) assert len(A__ ) == len(A__ ) return ref_ids def _A ( A__ ): """simple docstring""" with open(args.file_name , '''r''' , encoding='''utf-8''' ) as f: __lowercase = f.readlines() __lowercase = [line.strip() for line in data if len(A__ ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' __lowercase = LTP(args.ltp ) # faster in GPU device __lowercase = BertTokenizer.from_pretrained(args.bert ) __lowercase = prepare_ref(A__ , A__ , A__ ) with open(args.save_path , '''w''' , encoding='''utf-8''' ) as f: __lowercase = [json.dumps(A__ ) + '''\n''' for ref in ref_ids] f.writelines(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser(description='''prepare_chinese_ref''') parser.add_argument( '''--file_name''', type=str, default='''./resources/chinese-demo.txt''', help='''file need process, same as training data in lm''', ) parser.add_argument( '''--ltp''', type=str, default='''./resources/ltp''', help='''resources for LTP tokenizer, usually a path''' ) parser.add_argument('''--bert''', type=str, default='''./resources/robert''', help='''resources for Bert tokenizer''') parser.add_argument('''--save_path''', type=str, default='''./resources/ref.txt''', help='''path to save res''') lowerCAmelCase__ = parser.parse_args() main(args)
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ = [ '''small''', '''small-base''', '''medium''', '''medium-base''', '''intermediate''', '''intermediate-base''', '''large''', '''large-base''', '''xlarge''', '''xlarge-base''', ] lowerCAmelCase__ = { '''vocab_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json''', '''funnel-transformer/small-base''': ( '''https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json''', '''funnel-transformer/large-base''': ( '''https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase__ = {f'funnel-transformer/{name}': 512 for name in _model_names} lowerCAmelCase__ = {f'funnel-transformer/{name}': {'''do_lower_case''': True} for name in _model_names} class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : int = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : List[str] = FunnelTokenizer SCREAMING_SNAKE_CASE : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : int = 2 def __init__( self : str ,lowercase__ : List[str]=None ,lowercase__ : Dict=None ,lowercase__ : Optional[int]=True ,lowercase__ : Optional[int]="<unk>" ,lowercase__ : List[Any]="<sep>" ,lowercase__ : List[Any]="<pad>" ,lowercase__ : str="<cls>" ,lowercase__ : List[str]="<mask>" ,lowercase__ : Any="<s>" ,lowercase__ : str="</s>" ,lowercase__ : int=True ,lowercase__ : Dict=True ,lowercase__ : int=None ,lowercase__ : Dict="##" ,**lowercase__ : Optional[Any] ,): super().__init__( lowercase__ ,tokenizer_file=lowercase__ ,do_lower_case=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,clean_text=lowercase__ ,tokenize_chinese_chars=lowercase__ ,strip_accents=lowercase__ ,wordpieces_prefix=lowercase__ ,**lowercase__ ,) __lowercase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' ,lowercase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' ,lowercase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' ,lowercase__ ) != tokenize_chinese_chars ): __lowercase = getattr(lowercase__ ,normalizer_state.pop('''type''' ) ) __lowercase = do_lower_case __lowercase = strip_accents __lowercase = tokenize_chinese_chars __lowercase = normalizer_class(**lowercase__ ) __lowercase = do_lower_case def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[int] ,lowercase__ : int=None ): __lowercase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): __lowercase = self._tokenizer.model.save(lowercase__ ,name=lowercase__ ) return tuple(lowercase__ )
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib lowerCAmelCase__ = { '''debug''': logging.DEBUG, '''info''': logging.INFO, '''warning''': logging.WARNING, '''error''': logging.ERROR, '''critical''': logging.CRITICAL, } lowerCAmelCase__ = logging.WARNING def _A ( ): """simple docstring""" __lowercase = os.getenv('''DATASETS_VERBOSITY''' , A__ ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"Unknown option DATASETS_VERBOSITY={env_level_str}, " F"has to be one of: { ', '.join(log_levels.keys() ) }" ) return _default_log_level def _A ( ): """simple docstring""" return __name__.split('''.''' )[0] def _A ( ): """simple docstring""" return logging.getLogger(_get_library_name() ) def _A ( ): """simple docstring""" __lowercase = _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level() ) def _A ( ): """simple docstring""" __lowercase = _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET ) def _A ( A__ = None ): """simple docstring""" if name is None: __lowercase = _get_library_name() return logging.getLogger(A__ ) def _A ( ): """simple docstring""" return _get_library_root_logger().getEffectiveLevel() def _A ( A__ ): """simple docstring""" _get_library_root_logger().setLevel(A__ ) def _A ( ): """simple docstring""" return set_verbosity(A__ ) def _A ( ): """simple docstring""" return set_verbosity(A__ ) def _A ( ): """simple docstring""" return set_verbosity(A__ ) def _A ( ): """simple docstring""" return set_verbosity(A__ ) def _A ( ): """simple docstring""" __lowercase = False def _A ( ): """simple docstring""" __lowercase = True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class lowercase_ : def __init__( self : Optional[int] ,*lowercase__ : Any ,**lowercase__ : int ): # pylint: disable=unused-argument __lowercase = args[0] if args else None def __iter__( self : Any ): return iter(self._iterator ) def __getattr__( self : Union[str, Any] ,lowercase__ : List[str] ): def empty_fn(*lowercase__ : Dict ,**lowercase__ : int ): # pylint: disable=unused-argument return return empty_fn def __enter__( self : Union[str, Any] ): return self def __exit__( self : str ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : List[str] ): return lowerCAmelCase__ = True class lowercase_ : def __call__( self : Optional[int] ,*lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=False ,**lowercase__ : Optional[Any] ): if _tqdm_active and not disable: return tqdm_lib.tqdm(*lowercase__ ,**lowercase__ ) else: return EmptyTqdm(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,*lowercase__ : Any ,**lowercase__ : str ): __lowercase = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): if _tqdm_active: return tqdm_lib.tqdm.get_lock() lowerCAmelCase__ = _tqdm_cls() def _A ( ): """simple docstring""" global _tqdm_active return bool(_tqdm_active ) def _A ( ): """simple docstring""" global _tqdm_active __lowercase = True def _A ( ): """simple docstring""" global _tqdm_active __lowercase = False
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''vinvino02/glpn-kitti''': '''https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json''', # See all GLPN models at https://huggingface.co/models?filter=glpn } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = 'glpn' def __init__( self : Any ,lowercase__ : Optional[int]=3 ,lowercase__ : Any=4 ,lowercase__ : Optional[int]=[2, 2, 2, 2] ,lowercase__ : List[Any]=[8, 4, 2, 1] ,lowercase__ : Any=[3_2, 6_4, 1_6_0, 2_5_6] ,lowercase__ : Optional[Any]=[7, 3, 3, 3] ,lowercase__ : List[Any]=[4, 2, 2, 2] ,lowercase__ : List[Any]=[1, 2, 5, 8] ,lowercase__ : Optional[Any]=[4, 4, 4, 4] ,lowercase__ : str="gelu" ,lowercase__ : Any=0.0 ,lowercase__ : Tuple=0.0 ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : Dict=0.1 ,lowercase__ : List[Any]=1e-6 ,lowercase__ : int=6_4 ,lowercase__ : str=1_0 ,lowercase__ : str=-1 ,**lowercase__ : int ,): super().__init__(**lowercase__ ) __lowercase = num_channels __lowercase = num_encoder_blocks __lowercase = depths __lowercase = sr_ratios __lowercase = hidden_sizes __lowercase = patch_sizes __lowercase = strides __lowercase = mlp_ratios __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = drop_path_rate __lowercase = layer_norm_eps __lowercase = decoder_hidden_size __lowercase = max_depth __lowercase = head_in_index
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" if number > 0: raise ValueError('''input must be a negative integer''' ) __lowercase = len(bin(A__ )[3:] ) __lowercase = bin(abs(A__ ) - (1 << binary_number_length) )[3:] __lowercase = ( ( '''1''' + '''0''' * (binary_number_length - len(A__ )) + twos_complement_number ) if number < 0 else '''0''' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ).convert('''RGB''' ) __lowercase = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3) , (0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1) ), ] ) __lowercase = transform(A__ ).unsqueeze(0 ).to(A__ ) return image def _A ( A__ ): """simple docstring""" if "visual_encoder" in key: __lowercase = re.sub('''visual_encoder*''' , '''vision_model.encoder''' , A__ ) if "blocks" in key: __lowercase = re.sub(R'''blocks''' , '''layers''' , A__ ) if "attn" in key: __lowercase = re.sub(R'''attn''' , '''self_attn''' , A__ ) if "norm1" in key: __lowercase = re.sub(R'''norm1''' , '''layer_norm1''' , A__ ) if "norm2" in key: __lowercase = re.sub(R'''norm2''' , '''layer_norm2''' , A__ ) if "encoder.norm" in key: __lowercase = re.sub(R'''encoder.norm''' , '''post_layernorm''' , A__ ) if "encoder.patch_embed.proj" in key: __lowercase = re.sub(R'''encoder.patch_embed.proj''' , '''embeddings.patch_embedding''' , A__ ) if "encoder.pos_embed" in key: __lowercase = re.sub(R'''encoder.pos_embed''' , '''embeddings.position_embedding''' , A__ ) if "encoder.cls_token" in key: __lowercase = re.sub(R'''encoder.cls_token''' , '''embeddings.class_embedding''' , A__ ) if "self_attn" in key: __lowercase = re.sub(R'''self_attn.proj''' , '''self_attn.projection''' , A__ ) return key @torch.no_grad() def _A ( A__ , A__=None ): """simple docstring""" if config_path is not None: __lowercase = BlipConfig.from_pretrained(A__ ) else: __lowercase = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) __lowercase = BlipForConditionalGeneration(A__ ).eval() __lowercase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth''' __lowercase = blip_decoder(pretrained=A__ , image_size=384 , vit='''base''' ) __lowercase = pt_model.eval() __lowercase = pt_model.state_dict() for key in modified_state_dict.copy(): __lowercase = modified_state_dict.pop(A__ ) __lowercase = rename_key(A__ ) __lowercase = value hf_model.load_state_dict(A__ ) __lowercase = 384 __lowercase = load_demo_image(image_size=A__ , device='''cpu''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) __lowercase = tokenizer(['''a picture of'''] ).input_ids __lowercase = hf_model.generate(A__ , A__ ) assert out[0].tolist() == [30522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] __lowercase = hf_model.generate(A__ ) assert out[0].tolist() == [30522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(A__ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' __lowercase = ( '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth''' ) __lowercase = blip_vqa(pretrained=A__ , image_size=A__ , vit='''base''' ) vqa_model.eval() __lowercase = vqa_model.state_dict() for key in modified_state_dict.copy(): __lowercase = modified_state_dict.pop(A__ ) __lowercase = rename_key(A__ ) __lowercase = value __lowercase = BlipForQuestionAnswering(A__ ) hf_vqa_model.load_state_dict(A__ ) __lowercase = ['''How many dogs are in this image?'''] __lowercase = tokenizer(A__ , return_tensors='''pt''' ).input_ids __lowercase = hf_vqa_model.generate(A__ , A__ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '''_vqa''' ) __lowercase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth''' __lowercase = blip_itm(pretrained=A__ , image_size=A__ , vit='''base''' ) itm_model.eval() __lowercase = itm_model.state_dict() for key in modified_state_dict.copy(): __lowercase = modified_state_dict.pop(A__ ) __lowercase = rename_key(A__ ) __lowercase = value __lowercase = BlipForImageTextRetrieval(A__ ) __lowercase = ['''A picture of a woman with a dog sitting in a beach'''] __lowercase = tokenizer( A__ , return_tensors='''pt''' , padding='''max_length''' , truncation=A__ , max_length=35 , ).input_ids hf_itm_model.load_state_dict(A__ ) hf_itm_model.eval() __lowercase = hf_itm_model(A__ , A__ , use_itm_head=A__ ) __lowercase = hf_itm_model(A__ , A__ , use_itm_head=A__ ) assert out[0].item() == 0.2_1_1_0_6_8_7_4_9_4_2_7_7_9_5_4 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_5_6_9_8_8_4_5_3_8_6_5_0_5_1_2_7 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '''_itm''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCAmelCase__ = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
import os from collections import deque import torch from torch.utils.data import Dataset class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Dict="" ,lowercase__ : Optional[int]="train" ): assert os.path.isdir(lowercase__ ) __lowercase = [] __lowercase = os.listdir(lowercase__ ) for story_filename in story_filenames_list: if "summary" in story_filename: continue __lowercase = os.path.join(lowercase__ ,lowercase__ ) if not os.path.isfile(lowercase__ ): continue self.documents.append(lowercase__ ) def __len__( self : Optional[int] ): return len(self.documents ) def __getitem__( self : Optional[Any] ,lowercase__ : Tuple ): __lowercase = self.documents[idx] __lowercase = document_path.split('''/''' )[-1] with open(lowercase__ ,encoding='''utf-8''' ) as source: __lowercase = source.read() __lowercase , __lowercase = process_story(lowercase__ ) return document_name, story_lines, summary_lines def _A ( A__ ): """simple docstring""" __lowercase = list(filter(lambda A__ : len(A__ ) != 0 , [line.strip() for line in raw_story.split('''\n''' )] ) ) # for some unknown reason some lines miss a period, add it __lowercase = [_add_missing_period(A__ ) for line in nonempty_lines] # gather article lines __lowercase = [] __lowercase = deque(A__ ) while True: try: __lowercase = lines.popleft() if element.startswith('''@highlight''' ): break story_lines.append(A__ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines __lowercase = list(filter(lambda A__ : not t.startswith('''@highlight''' ) , A__ ) ) return story_lines, summary_lines def _A ( A__ ): """simple docstring""" __lowercase = ['''.''', '''!''', '''?''', '''...''', '''\'''', '''`''', '''"''', '''\u2019''', '''\u2019''', ''')'''] if line.startswith('''@highlight''' ): return line if line[-1] in END_TOKENS: return line return line + "." def _A ( A__ , A__ , A__ ): """simple docstring""" if len(A__ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(A__ )) ) return sequence def _A ( A__ , A__ ): """simple docstring""" __lowercase = torch.ones_like(A__ ) __lowercase = sequence == pad_token_id __lowercase = 0 return mask def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = [tokenizer.encode(A__ ) for line in story_lines] __lowercase = [token for sentence in story_lines_token_ids for token in sentence] __lowercase = [tokenizer.encode(A__ ) for line in summary_lines] __lowercase = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] for sequence in batch: __lowercase = -1 __lowercase = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(A__ ) return torch.tensor(A__ )
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class lowercase_ : """simple docstring""" def __init__( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=1_3 ,lowercase__ : Optional[int]=3_2 ,lowercase__ : Optional[Any]=2 ,lowercase__ : Union[str, Any]=3 ,lowercase__ : Any=1_6 ,lowercase__ : List[str]=[1, 2, 1] ,lowercase__ : List[str]=[2, 2, 4] ,lowercase__ : Optional[Any]=2 ,lowercase__ : Union[str, Any]=2.0 ,lowercase__ : Optional[int]=True ,lowercase__ : Tuple=0.0 ,lowercase__ : int=0.0 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : List[Any]="gelu" ,lowercase__ : Optional[Any]=False ,lowercase__ : List[Any]=True ,lowercase__ : Dict=0.0_2 ,lowercase__ : Union[str, Any]=1e-5 ,lowercase__ : List[str]=True ,lowercase__ : Optional[int]=None ,lowercase__ : str=True ,lowercase__ : Optional[int]=1_0 ,lowercase__ : Dict=8 ,lowercase__ : Optional[int]=["stage1", "stage2", "stage3"] ,lowercase__ : Dict=[1, 2, 3] ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return MaskFormerSwinConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,depths=self.depths ,num_heads=self.num_heads ,window_size=self.window_size ,mlp_ratio=self.mlp_ratio ,qkv_bias=self.qkv_bias ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,drop_path_rate=self.drop_path_rate ,hidden_act=self.hidden_act ,use_absolute_embeddings=self.use_absolute_embeddings ,path_norm=self.patch_norm ,layer_norm_eps=self.layer_norm_eps ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,out_features=self.out_features ,out_indices=self.out_indices ,) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ): __lowercase = MaskFormerSwinModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, expected_seq_len, expected_dim) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ): __lowercase = MaskFormerSwinBackbone(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) ,len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) ,[1_3, 1_6, 1_6, 1_6] ) # verify channels self.parent.assertEqual(len(model.channels ) ,len(config.out_features ) ) self.parent.assertListEqual(model.channels ,[1_6, 3_2, 6_4] ) # verify ValueError with self.parent.assertRaises(lowercase__ ): __lowercase = ['''stem'''] __lowercase = MaskFormerSwinBackbone(config=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Dict = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : Any = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Dict = False def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,embed_dim=3_7 ) @require_torch_multi_gpu @unittest.skip( reason=( '''`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with''' ''' `nn.DataParallel`''' ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : Any ): return def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase__ ) @unittest.skip('''Swin does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass @unittest.skip('''Swin does not support feedforward chunking''' ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): pass def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) @unittest.skip(reason='''MaskFormerSwin is only used as backbone and doesn\'t support output_attentions''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass @unittest.skip(reason='''MaskFormerSwin is only used as an internal backbone''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : int ,lowercase__ : Union[str, Any] ,lowercase__ : str ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester ,'''expected_num_hidden_layers''' ,len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ,(padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ,(padded_height, padded_width) ) @unittest.skip(reason='''MaskFormerSwin doesn\'t have pretrained checkpoints''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass @unittest.skip(reason='''This will be fixed once MaskFormerSwin is replaced by native Swin''' ) def SCREAMING_SNAKE_CASE ( self : int ): pass @unittest.skip(reason='''This will be fixed once MaskFormerSwin is replaced by native Swin''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(lowercase__ : Union[str, Any] ): __lowercase = 0 return t def check_equivalence(lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : str ,lowercase__ : Union[str, Any]={} ): with torch.no_grad(): __lowercase = model(**lowercase__ ,return_dict=lowercase__ ,**lowercase__ ) __lowercase = model(**lowercase__ ,return_dict=lowercase__ ,**lowercase__ ).to_tuple() def recursive_check(lowercase__ : Optional[int] ,lowercase__ : Optional[Any] ): if isinstance(lowercase__ ,(List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowercase__ ,lowercase__ ): recursive_check(lowercase__ ,lowercase__ ) elif isinstance(lowercase__ ,lowercase__ ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() ,dict_object.values() ): recursive_check(lowercase__ ,lowercase__ ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(lowercase__ ) ,set_nan_tensor_to_zero(lowercase__ ) ,atol=1e-5 ) ,msg=( '''Tuple and dict output are not equal. Difference:''' F" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" F" {torch.isnan(lowercase__ ).any()} and `inf`: {torch.isinf(lowercase__ )}. Dict has" F" `nan`: {torch.isnan(lowercase__ ).any()} and `inf`: {torch.isinf(lowercase__ )}." ) ,) recursive_check(lowercase__ ,lowercase__ ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) check_equivalence(lowercase__ ,lowercase__ ,lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) check_equivalence(lowercase__ ,lowercase__ ,lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ) check_equivalence(lowercase__ ,lowercase__ ,lowercase__ ,{'''output_hidden_states''': True} ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) check_equivalence(lowercase__ ,lowercase__ ,lowercase__ ,{'''output_hidden_states''': True} ) @require_torch class lowercase_ (unittest.TestCase , lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = (MaskFormerSwinBackbone,) if is_torch_available() else () SCREAMING_SNAKE_CASE : Dict = MaskFormerSwinConfig def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = MaskFormerSwinModelTester(self ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict['''pixel_values'''].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(lowercase__ ) backbone.to(lowercase__ ) backbone.eval() __lowercase = backbone(**lowercase__ ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps ,lowercase__ ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps ,backbone.channels ): self.assertTrue(feature_map.shape[:2] ,(batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**lowercase__ ,output_hidden_states=lowercase__ ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) ,len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] ,backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) ,(batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**lowercase__ ,output_attentions=lowercase__ ) self.assertIsNotNone(outputs.attentions )
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ ): """simple docstring""" __lowercase = '''huggingface/label-files''' __lowercase = '''imagenet-1k-id2label.json''' __lowercase = json.load(open(hf_hub_download(A__ , A__ , repo_type='''dataset''' ) , '''r''' ) ) __lowercase = {int(A__ ): v for k, v in idalabel.items()} __lowercase = {v: k for k, v in idalabel.items()} __lowercase = '''std_conv''' if '''bit''' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" __lowercase = BitConfig( conv_layer=A__ , num_labels=1000 , idalabel=A__ , labelaid=A__ , ) return config def _A ( A__ ): """simple docstring""" if "stem.conv" in name: __lowercase = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: __lowercase = name.replace('''blocks''' , '''layers''' ) if "head.fc" in name: __lowercase = name.replace('''head.fc''' , '''classifier.1''' ) if name.startswith('''norm''' ): __lowercase = '''bit.''' + name if "bit" not in name and "classifier" not in name: __lowercase = '''bit.encoder.''' + name return name def _A ( ): """simple docstring""" __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def _A ( A__ , A__ , A__=False ): """simple docstring""" __lowercase = get_config(A__ ) # load original model from timm __lowercase = create_model(A__ , pretrained=A__ ) timm_model.eval() # load state_dict of original model __lowercase = timm_model.state_dict() for key in state_dict.copy().keys(): __lowercase = state_dict.pop(A__ ) __lowercase = val.squeeze() if '''head''' in key else val # load HuggingFace model __lowercase = BitForImageClassification(A__ ) model.eval() model.load_state_dict(A__ ) # create image processor __lowercase = create_transform(**resolve_data_config({} , model=A__ ) ) __lowercase = transform.transforms __lowercase = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } __lowercase = BitImageProcessor( do_resize=A__ , size={'''shortest_edge''': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=A__ , crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]} , do_normalize=A__ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) __lowercase = prepare_img() __lowercase = transform(A__ ).unsqueeze(0 ) __lowercase = processor(A__ , return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(A__ , A__ ) # verify logits with torch.no_grad(): __lowercase = model(A__ ) __lowercase = outputs.logits print('''Logits:''' , logits[0, :3] ) print('''Predicted class:''' , model.config.idalabel[logits.argmax(-1 ).item()] ) __lowercase = timm_model(A__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(A__ , outputs.logits , atol=1e-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(A__ ).mkdir(exist_ok=A__ ) print(F"Saving model {model_name} and processor to {pytorch_dump_folder_path}" ) model.save_pretrained(A__ ) processor.save_pretrained(A__ ) if push_to_hub: print(F"Pushing model {model_name} and processor to the hub" ) model.push_to_hub(F"ybelkada/{model_name}" ) processor.push_to_hub(F"ybelkada/{model_name}" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''resnetv2_50x1_bitm''', type=str, help='''Name of the BiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model to the hub.''', ) lowerCAmelCase__ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCAmelCase__ = {'''configuration_yolos''': ['''YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''YolosConfig''', '''YolosOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''YolosFeatureExtractor'''] lowerCAmelCase__ = ['''YolosImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''YolosForObjectDetection''', '''YolosModel''', '''YolosPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCAmelCase__ = logging.get_logger(__name__) @add_end_docstrings(lowerCamelCase__ ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : List[str] ,*lowercase__ : List[str] ,**lowercase__ : str ): super().__init__(*lowercase__ ,**lowercase__ ) requires_backends(self ,'''vision''' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Optional[Any]=None ): __lowercase = {} if top_k is not None: __lowercase = top_k return {}, {}, postprocess_params def __call__( self : int ,lowercase__ : Union[str, List[str], "Image.Image", List["Image.Image"]] ,**lowercase__ : Dict ): return super().__call__(lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : int ): __lowercase = load_image(lowercase__ ) __lowercase = self.image_processor(images=lowercase__ ,return_tensors=self.framework ) return model_inputs def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Optional[int] ): __lowercase = self.model(**lowercase__ ) return model_outputs def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Dict ,lowercase__ : List[Any]=5 ): if top_k > self.model.config.num_labels: __lowercase = self.model.config.num_labels if self.framework == "pt": __lowercase = model_outputs.logits.softmax(-1 )[0] __lowercase , __lowercase = probs.topk(lowercase__ ) elif self.framework == "tf": __lowercase = stable_softmax(model_outputs.logits ,axis=-1 )[0] __lowercase = tf.math.top_k(lowercase__ ,k=lowercase__ ) __lowercase , __lowercase = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(F"Unsupported framework: {self.framework}" ) __lowercase = scores.tolist() __lowercase = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(lowercase__ ,lowercase__ )]
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP lowerCAmelCase__ = False try: lowerCAmelCase__ = _is_package_available('''google.colab''') except ModuleNotFoundError: pass @input.register class lowercase_ : """simple docstring""" def __init__( self : str ,lowercase__ : str = None ,lowercase__ : list = [] ): __lowercase = 0 __lowercase = choices __lowercase = prompt if sys.platform == "win32": __lowercase = '''*''' else: __lowercase = '''➔ ''' def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : int ,lowercase__ : str = "" ): if sys.platform != "win32": writeColor(self.choices[index] ,3_2 ,lowercase__ ) else: forceWrite(self.choices[index] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : int ): if index == self.position: forceWrite(F" {self.arrow_char} " ) self.write_choice(lowercase__ ) else: forceWrite(F" {self.choices[index]}" ) reset_cursor() def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Direction ,lowercase__ : int = 1 ): __lowercase = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(lowercase__ ) move_cursor(lowercase__ ,direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP['''up'''] ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.move_direction(Direction.UP ) @input.mark(KEYMAP['''down'''] ) def SCREAMING_SNAKE_CASE ( self : List[str] ): self.move_direction(Direction.DOWN ) @input.mark(KEYMAP['''newline'''] ) def SCREAMING_SNAKE_CASE ( self : str ): move_cursor(len(self.choices ) - self.position ,'''DOWN''' ) return self.position @input.mark(KEYMAP['''interrupt'''] ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): move_cursor(len(self.choices ) - self.position ,'''DOWN''' ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(lowercase__ )] for number in range(1_0 )] ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = int(chr(self.current_selection ) ) __lowercase = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP ,-movement ) elif self.position < index: self.move_direction(Direction.DOWN ,lowercase__ ) else: return else: return def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : int = 0 ): if self.prompt: linebreak() forceWrite(self.prompt ,'''\n''' ) if in_colab: forceWrite('''Please input a choice index (starting from 0), and press enter''' ,'''\n''' ) else: forceWrite('''Please select a choice using the arrow or number keys, and selecting with enter''' ,'''\n''' ) __lowercase = default_choice for i in range(len(self.choices ) ): self.print_choice(lowercase__ ) forceWrite('''\n''' ) move_cursor(len(self.choices ) - self.position ,'''UP''' ) with cursor.hide(): while True: if in_colab: try: __lowercase = int(builtins.input() ) except ValueError: __lowercase = default_choice else: __lowercase = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 ,'''UP''' ) clear_line() self.write_choice(lowercase__ ,'''\n''' ) return choice
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Tuple ,lowercase__ : List[str]=1_3 ,lowercase__ : str=7 ,lowercase__ : Optional[Any]=False ,lowercase__ : List[str]=True ,lowercase__ : Any=False ,lowercase__ : List[str]=True ,lowercase__ : Union[str, Any]=3_3 ,lowercase__ : str=3_2 ,lowercase__ : Any=5 ,lowercase__ : str=4 ,lowercase__ : List[Any]=3_7 ,lowercase__ : Tuple="gelu" ,lowercase__ : str=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : List[str]=5_1_2 ,lowercase__ : Tuple=1_6 ,lowercase__ : Dict=2 ,lowercase__ : Dict=0.0_2 ,lowercase__ : Optional[Any]=3 ,lowercase__ : List[str]=4 ,lowercase__ : Tuple=None ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Tuple ): return EsmConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,pad_token_id=1 ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : Any ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any] ,lowercase__ : str ): __lowercase = EsmModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ) __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : str ,lowercase__ : Any ,lowercase__ : Tuple ): __lowercase = EsmForMaskedLM(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Dict ,lowercase__ : Optional[Any] ,lowercase__ : Any ,lowercase__ : Tuple ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ): __lowercase = self.num_labels __lowercase = EsmForTokenClassification(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Union[str, Any] = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Any = () SCREAMING_SNAKE_CASE : List[Any] = ( { 'feature-extraction': EsmModel, 'fill-mask': EsmForMaskedLM, 'text-classification': EsmForSequenceClassification, 'token-classification': EsmForTokenClassification, 'zero-shot': EsmForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : int = True def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = EsmModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Dict ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowercase = type self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = EsmModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs()[0] __lowercase = EsmEmbeddings(config=lowercase__ ) __lowercase = torch.as_tensor([[1_2, 3_1, 1_3, model.padding_idx]] ) __lowercase = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) __lowercase = create_position_ids_from_input_ids(lowercase__ ,model.padding_idx ) self.assertEqual(position_ids.shape ,expected_positions.shape ) self.assertTrue(torch.all(torch.eq(lowercase__ ,lowercase__ ) ) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs()[0] __lowercase = EsmEmbeddings(config=lowercase__ ) __lowercase = torch.empty(2 ,4 ,3_0 ) __lowercase = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] __lowercase = torch.as_tensor([expected_single_positions, expected_single_positions] ) __lowercase = embeddings.create_position_ids_from_inputs_embeds(lowercase__ ) self.assertEqual(position_ids.shape ,expected_positions.shape ) self.assertTrue(torch.all(torch.eq(lowercase__ ,lowercase__ ) ) ) @unittest.skip('''Esm does not support embedding resizing''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): pass @unittest.skip('''Esm does not support embedding resizing''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @require_torch class lowercase_ (lowerCamelCase__ ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): with torch.no_grad(): __lowercase = EsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) model.eval() __lowercase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __lowercase = model(lowercase__ )[0] __lowercase = 3_3 __lowercase = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape ,lowercase__ ) __lowercase = torch.tensor( [[[8.9_2_1_5, -1_0.5_8_9_8, -6.4_6_7_1], [-6.3_9_6_7, -1_3.9_1_1_4, -1.1_2_1_2], [-7.7_8_1_2, -1_3.9_5_1_6, -3.7_4_0_6]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,lowercase__ ,atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): with torch.no_grad(): __lowercase = EsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) model.eval() __lowercase = torch.tensor([[0, 6, 4, 1_3, 5, 4, 1_6, 1_2, 1_1, 7, 2]] ) __lowercase = model(lowercase__ )[0] # compare the actual values for a slice. __lowercase = torch.tensor( [[[0.1_4_4_4, 0.5_4_1_3, 0.3_2_4_8], [0.3_0_3_4, 0.0_0_5_3, 0.3_1_0_8], [0.3_2_2_8, -0.2_4_9_9, 0.3_4_1_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,lowercase__ ,atol=1e-4 ) )
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter lowerCAmelCase__ = True except ImportError: lowerCAmelCase__ = False lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name def _A ( A__ ): """simple docstring""" return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''add-new-model''' ) add_new_model_parser.add_argument('''--testing''' ,action='''store_true''' ,help='''If in testing mode.''' ) add_new_model_parser.add_argument('''--testing_file''' ,type=lowercase__ ,help='''Configuration file on which to run.''' ) add_new_model_parser.add_argument( '''--path''' ,type=lowercase__ ,help='''Path to cookiecutter. Should only be used for testing purposes.''' ) add_new_model_parser.set_defaults(func=lowercase__ ) def __init__( self : List[Any] ,lowercase__ : bool ,lowercase__ : str ,lowercase__ : Tuple=None ,*lowercase__ : int ): __lowercase = testing __lowercase = testing_file __lowercase = path def SCREAMING_SNAKE_CASE ( self : List[str] ): warnings.warn( '''The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ''' '''It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ''' '''checks, you should use `transformers-cli add-new-model-like` instead.''' ) if not _has_cookiecutter: raise ImportError( '''Model creation dependencies are required to use the `add_new_model` command. Install them by running ''' '''the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n''' ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory __lowercase = [directory for directory in os.listdir() if '''cookiecutter-template-''' == directory[:2_2]] if len(lowercase__ ) > 0: raise ValueError( '''Several directories starting with `cookiecutter-template-` in current working directory. ''' '''Please clean your directory by removing all folders starting with `cookiecutter-template-` or ''' '''change your working directory.''' ) __lowercase = ( Path(lowercase__ ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) __lowercase = path_to_transformer_root / '''templates''' / '''adding_a_new_model''' # Execute cookiecutter if not self._testing: cookiecutter(str(lowercase__ ) ) else: with open(self._testing_file ,'''r''' ) as configuration_file: __lowercase = json.load(lowercase__ ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) ,no_input=lowercase__ ,extra_context=lowercase__ ,) __lowercase = [directory for directory in os.listdir() if '''cookiecutter-template-''' in directory[:2_2]][0] # Retrieve configuration with open(directory + '''/configuration.json''' ,'''r''' ) as configuration_file: __lowercase = json.load(lowercase__ ) __lowercase = configuration['''lowercase_modelname'''] __lowercase = configuration['''generate_tensorflow_pytorch_and_flax'''] os.remove(F"{directory}/configuration.json" ) __lowercase = '''PyTorch''' in generate_tensorflow_pytorch_and_flax __lowercase = '''TensorFlow''' in generate_tensorflow_pytorch_and_flax __lowercase = '''Flax''' in generate_tensorflow_pytorch_and_flax __lowercase = F"{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}" os.makedirs(lowercase__ ,exist_ok=lowercase__ ) os.makedirs(F"{path_to_transformer_root}/tests/models/{lowercase_model_name}" ,exist_ok=lowercase__ ) # Tests require submodules as they have parent imports with open(F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py" ,'''w''' ): pass shutil.move( F"{directory}/__init__.py" ,F"{model_dir}/__init__.py" ,) shutil.move( F"{directory}/configuration_{lowercase_model_name}.py" ,F"{model_dir}/configuration_{lowercase_model_name}.py" ,) def remove_copy_lines(lowercase__ : Optional[int] ): with open(lowercase__ ,'''r''' ) as f: __lowercase = f.readlines() with open(lowercase__ ,'''w''' ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(lowercase__ ) if output_pytorch: if not self._testing: remove_copy_lines(F"{directory}/modeling_{lowercase_model_name}.py" ) shutil.move( F"{directory}/modeling_{lowercase_model_name}.py" ,F"{model_dir}/modeling_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/test_modeling_{lowercase_model_name}.py" ,F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py" ,) else: os.remove(F"{directory}/modeling_{lowercase_model_name}.py" ) os.remove(F"{directory}/test_modeling_{lowercase_model_name}.py" ) if output_tensorflow: if not self._testing: remove_copy_lines(F"{directory}/modeling_tf_{lowercase_model_name}.py" ) shutil.move( F"{directory}/modeling_tf_{lowercase_model_name}.py" ,F"{model_dir}/modeling_tf_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/test_modeling_tf_{lowercase_model_name}.py" ,F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py" ,) else: os.remove(F"{directory}/modeling_tf_{lowercase_model_name}.py" ) os.remove(F"{directory}/test_modeling_tf_{lowercase_model_name}.py" ) if output_flax: if not self._testing: remove_copy_lines(F"{directory}/modeling_flax_{lowercase_model_name}.py" ) shutil.move( F"{directory}/modeling_flax_{lowercase_model_name}.py" ,F"{model_dir}/modeling_flax_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/test_modeling_flax_{lowercase_model_name}.py" ,F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py" ,) else: os.remove(F"{directory}/modeling_flax_{lowercase_model_name}.py" ) os.remove(F"{directory}/test_modeling_flax_{lowercase_model_name}.py" ) shutil.move( F"{directory}/{lowercase_model_name}.md" ,F"{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md" ,) shutil.move( F"{directory}/tokenization_{lowercase_model_name}.py" ,F"{model_dir}/tokenization_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/tokenization_fast_{lowercase_model_name}.py" ,F"{model_dir}/tokenization_{lowercase_model_name}_fast.py" ,) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(lowercase__ : str ,lowercase__ : str ,lowercase__ : List[str] ): # Create temp file __lowercase , __lowercase = mkstemp() __lowercase = False with fdopen(lowercase__ ,'''w''' ) as new_file: with open(lowercase__ ) as old_file: for line in old_file: new_file.write(lowercase__ ) if line_to_copy_below in line: __lowercase = True for line_to_copy in lines_to_copy: new_file.write(lowercase__ ) if not line_found: raise ValueError(F"Line {line_to_copy_below} was not found in file." ) # Copy the file permissions from the old file to the new file copymode(lowercase__ ,lowercase__ ) # Remove original file remove(lowercase__ ) # Move new file move(lowercase__ ,lowercase__ ) def skip_units(lowercase__ : Union[str, Any] ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(lowercase__ : Any ): with open(lowercase__ ) as datafile: __lowercase = [] __lowercase = False __lowercase = False for line in datafile: if "# To replace in: " in line and "##" not in line: __lowercase = line.split('''"''' )[1] __lowercase = skip_units(lowercase__ ) elif "# Below: " in line and "##" not in line: __lowercase = line.split('''"''' )[1] __lowercase = skip_units(lowercase__ ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(lowercase__ ,lowercase__ ,lowercase__ ) __lowercase = [] elif "# Replace with" in line and "##" not in line: __lowercase = [] elif "##" not in line: lines_to_copy.append(lowercase__ ) remove(lowercase__ ) replace_in_files(F"{directory}/to_replace_{lowercase_model_name}.py" ) os.rmdir(lowercase__ )
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ ): """simple docstring""" __lowercase = SwinConfig( embed_dim=192 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=['''stage2''', '''stage3''', '''stage4'''] , ) __lowercase = DetaConfig( backbone_config=A__ , num_queries=900 , encoder_ffn_dim=2048 , decoder_ffn_dim=2048 , num_feature_levels=5 , assign_first_stage=A__ , with_box_refine=A__ , two_stage=A__ , ) # set labels __lowercase = '''huggingface/label-files''' if "o365" in model_name: __lowercase = 366 __lowercase = '''object365-id2label.json''' else: __lowercase = 91 __lowercase = '''coco-detection-id2label.json''' __lowercase = num_labels __lowercase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type='''dataset''' ) ) , '''r''' ) ) __lowercase = {int(A__ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} return config def _A ( A__ ): """simple docstring""" __lowercase = [] # stem # fmt: off rename_keys.append(('''backbone.0.body.patch_embed.proj.weight''', '''model.backbone.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.0.body.patch_embed.proj.bias''', '''model.backbone.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.0.body.patch_embed.norm.weight''', '''model.backbone.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.0.body.patch_embed.norm.bias''', '''model.backbone.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm1.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm1.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm2.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.norm2.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias", F"model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias") ) if i < 3: rename_keys.append((F"backbone.0.body.layers.{i}.downsample.reduction.weight", F"model.backbone.model.encoder.layers.{i}.downsample.reduction.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.downsample.norm.weight", F"model.backbone.model.encoder.layers.{i}.downsample.norm.weight") ) rename_keys.append((F"backbone.0.body.layers.{i}.downsample.norm.bias", F"model.backbone.model.encoder.layers.{i}.downsample.norm.bias") ) rename_keys.append(('''backbone.0.body.norm1.weight''', '''model.backbone.model.hidden_states_norms.stage2.weight''') ) rename_keys.append(('''backbone.0.body.norm1.bias''', '''model.backbone.model.hidden_states_norms.stage2.bias''') ) rename_keys.append(('''backbone.0.body.norm2.weight''', '''model.backbone.model.hidden_states_norms.stage3.weight''') ) rename_keys.append(('''backbone.0.body.norm2.bias''', '''model.backbone.model.hidden_states_norms.stage3.bias''') ) rename_keys.append(('''backbone.0.body.norm3.weight''', '''model.backbone.model.hidden_states_norms.stage4.weight''') ) rename_keys.append(('''backbone.0.body.norm3.bias''', '''model.backbone.model.hidden_states_norms.stage4.bias''') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight", F"model.encoder.layers.{i}.self_attn.sampling_offsets.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias", F"model.encoder.layers.{i}.self_attn.sampling_offsets.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.attention_weights.weight", F"model.encoder.layers.{i}.self_attn.attention_weights.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.attention_weights.bias", F"model.encoder.layers.{i}.self_attn.attention_weights.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.value_proj.weight", F"model.encoder.layers.{i}.self_attn.value_proj.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.value_proj.bias", F"model.encoder.layers.{i}.self_attn.value_proj.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.output_proj.weight", F"model.encoder.layers.{i}.self_attn.output_proj.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.self_attn.output_proj.bias", F"model.encoder.layers.{i}.self_attn.output_proj.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm1.weight", F"model.encoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm1.bias", F"model.encoder.layers.{i}.self_attn_layer_norm.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear1.weight", F"model.encoder.layers.{i}.fc1.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear1.bias", F"model.encoder.layers.{i}.fc1.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear2.weight", F"model.encoder.layers.{i}.fc2.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear2.bias", F"model.encoder.layers.{i}.fc2.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm2.weight", F"model.encoder.layers.{i}.final_layer_norm.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm2.bias", F"model.encoder.layers.{i}.final_layer_norm.bias") ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight", F"model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias", F"model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.attention_weights.weight", F"model.decoder.layers.{i}.encoder_attn.attention_weights.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.attention_weights.bias", F"model.decoder.layers.{i}.encoder_attn.attention_weights.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.value_proj.weight", F"model.decoder.layers.{i}.encoder_attn.value_proj.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.value_proj.bias", F"model.decoder.layers.{i}.encoder_attn.value_proj.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.output_proj.weight", F"model.decoder.layers.{i}.encoder_attn.output_proj.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.cross_attn.output_proj.bias", F"model.decoder.layers.{i}.encoder_attn.output_proj.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm1.weight", F"model.decoder.layers.{i}.encoder_attn_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm1.bias", F"model.decoder.layers.{i}.encoder_attn_layer_norm.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.self_attn.out_proj.weight", F"model.decoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.self_attn.out_proj.bias", F"model.decoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm2.weight", F"model.decoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm2.bias", F"model.decoder.layers.{i}.self_attn_layer_norm.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear1.weight", F"model.decoder.layers.{i}.fc1.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear1.bias", F"model.decoder.layers.{i}.fc1.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear2.weight", F"model.decoder.layers.{i}.fc2.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.linear2.bias", F"model.decoder.layers.{i}.fc2.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm3.weight", F"model.decoder.layers.{i}.final_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm3.bias", F"model.decoder.layers.{i}.final_layer_norm.bias") ) # fmt: on return rename_keys def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = dct.pop(A__ ) __lowercase = val def _A ( A__ , A__ ): """simple docstring""" __lowercase = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __lowercase = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __lowercase = state_dict.pop(F"backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight" ) __lowercase = state_dict.pop(F"backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase = in_proj_weight[:dim, :] __lowercase = in_proj_bias[: dim] __lowercase = in_proj_weight[ dim : dim * 2, : ] __lowercase = in_proj_bias[ dim : dim * 2 ] __lowercase = in_proj_weight[ -dim :, : ] __lowercase = in_proj_bias[-dim :] # fmt: on def _A ( A__ , A__ ): """simple docstring""" __lowercase = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention __lowercase = state_dict.pop(F"transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) __lowercase = state_dict.pop(F"transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase = in_proj_weight[:hidden_size, :] __lowercase = in_proj_bias[:hidden_size] __lowercase = in_proj_weight[ hidden_size : hidden_size * 2, : ] __lowercase = in_proj_bias[hidden_size : hidden_size * 2] __lowercase = in_proj_weight[-hidden_size:, :] __lowercase = in_proj_bias[-hidden_size:] def _A ( ): """simple docstring""" __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = get_deta_config(A__ ) # load original state dict if model_name == "deta-swin-large": __lowercase = hf_hub_download(repo_id='''nielsr/deta-checkpoints''' , filename='''adet_swin_ft.pth''' ) elif model_name == "deta-swin-large-o365": __lowercase = hf_hub_download(repo_id='''jozhang97/deta-swin-l-o365''' , filename='''deta_swin_pt_o365.pth''' ) else: raise ValueError(F"Model name {model_name} not supported" ) __lowercase = torch.load(A__ , map_location='''cpu''' )['''model'''] # original state dict for name, param in state_dict.items(): print(A__ , param.shape ) # rename keys __lowercase = create_rename_keys(A__ ) for src, dest in rename_keys: rename_key(A__ , A__ , A__ ) read_in_swin_q_k_v(A__ , config.backbone_config ) read_in_decoder_q_k_v(A__ , A__ ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: __lowercase = state_dict.pop(A__ ) __lowercase = val if "input_proj" in key: __lowercase = state_dict.pop(A__ ) __lowercase = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: __lowercase = state_dict.pop(A__ ) __lowercase = val # finally, create HuggingFace model and load state dict __lowercase = DetaForObjectDetection(A__ ) model.load_state_dict(A__ ) model.eval() __lowercase = '''cuda''' if torch.cuda.is_available() else '''cpu''' model.to(A__ ) # load image processor __lowercase = DetaImageProcessor(format='''coco_detection''' ) # verify our conversion on image __lowercase = prepare_img() __lowercase = processor(images=A__ , return_tensors='''pt''' ) __lowercase = encoding['''pixel_values'''] __lowercase = model(pixel_values.to(A__ ) ) # verify logits print('''Logits:''' , outputs.logits[0, :3, :3] ) print('''Boxes:''' , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": __lowercase = torch.tensor( [[-7.6_3_0_8, -2.8_4_8_5, -5.3_7_3_7], [-7.2_0_3_7, -4.5_5_0_5, -4.8_0_2_7], [-7.2_9_4_3, -4.2_6_1_1, -4.6_6_1_7]] ) __lowercase = torch.tensor([[0.4_9_8_7, 0.4_9_6_9, 0.9_9_9_9], [0.2_5_4_9, 0.5_4_9_8, 0.4_8_0_5], [0.5_4_9_8, 0.2_7_5_7, 0.0_5_6_9]] ) elif model_name == "deta-swin-large-o365": __lowercase = torch.tensor( [[-8.0_1_2_2, -3.5_7_2_0, -4.9_7_1_7], [-8.1_5_4_7, -3.6_8_8_6, -4.6_3_8_9], [-7.6_6_1_0, -3.6_1_9_4, -5.0_1_3_4]] ) __lowercase = torch.tensor([[0.2_5_2_3, 0.5_5_4_9, 0.4_8_8_1], [0.7_7_1_5, 0.4_1_4_9, 0.4_6_0_1], [0.5_5_0_3, 0.2_7_5_3, 0.0_5_7_5]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(A__ ) , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(A__ ) , atol=1e-4 ) print('''Everything ok!''' ) if pytorch_dump_folder_path: # Save model and processor logger.info(F"Saving PyTorch model and processor to {pytorch_dump_folder_path}..." ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) processor.save_pretrained(A__ ) # Push to hub if push_to_hub: print('''Pushing model and processor to hub...''' ) model.push_to_hub(F"jozhang97/{model_name}" ) processor.push_to_hub(F"jozhang97/{model_name}" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', type=str, default='''deta-swin-large''', choices=['''deta-swin-large''', '''deta-swin-large-o365'''], help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCAmelCase__ = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : torch.FloatTensor SCREAMING_SNAKE_CASE : Optional[torch.FloatTensor] = None def _A ( A__ , A__=0.9_9_9 , A__="cosine" , ): """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(A__ ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(A__ ): return math.exp(t * -12.0 ) else: raise ValueError(F"Unsupported alpha_tranform_type: {alpha_transform_type}" ) __lowercase = [] for i in range(A__ ): __lowercase = i / num_diffusion_timesteps __lowercase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(A__ ) / alpha_bar_fn(A__ ) , A__ ) ) return torch.tensor(A__ , dtype=torch.floataa ) class lowercase_ (lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : str = 1 @register_to_config def __init__( self : Union[str, Any] ,lowercase__ : int = 1_0_0_0 ,lowercase__ : float = 0.0_0_0_1 ,lowercase__ : float = 0.0_2 ,lowercase__ : str = "linear" ,lowercase__ : Optional[Union[np.ndarray, List[float]]] = None ,lowercase__ : bool = True ,lowercase__ : bool = True ,lowercase__ : int = 0 ,lowercase__ : str = "epsilon" ,lowercase__ : float = 1.0 ,**lowercase__ : Optional[int] ,): if kwargs.get('''set_alpha_to_one''' ,lowercase__ ) is not None: __lowercase = ( '''The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead.''' ) deprecate('''set_alpha_to_one''' ,'''1.0.0''' ,lowercase__ ,standard_warn=lowercase__ ) __lowercase = kwargs['''set_alpha_to_one'''] if trained_betas is not None: __lowercase = torch.tensor(lowercase__ ,dtype=torch.floataa ) elif beta_schedule == "linear": __lowercase = torch.linspace(lowercase__ ,lowercase__ ,lowercase__ ,dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __lowercase = ( torch.linspace(beta_start**0.5 ,beta_end**0.5 ,lowercase__ ,dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __lowercase = betas_for_alpha_bar(lowercase__ ) else: raise NotImplementedError(F"{beta_schedule} does is not implemented for {self.__class__}" ) __lowercase = 1.0 - self.betas __lowercase = torch.cumprod(self.alphas ,dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. __lowercase = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution __lowercase = 1.0 # setable values __lowercase = None __lowercase = torch.from_numpy(np.arange(0 ,lowercase__ ).copy().astype(np.intaa ) ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : torch.FloatTensor ,lowercase__ : Optional[int] = None ): return sample def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : int ,lowercase__ : Union[str, torch.device] = None ): if num_inference_steps > self.config.num_train_timesteps: raise ValueError( F"`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:" F" {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle" F" maximal {self.config.num_train_timesteps} timesteps." ) __lowercase = num_inference_steps __lowercase = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __lowercase = (np.arange(0 ,lowercase__ ) * step_ratio).round().copy().astype(np.intaa ) __lowercase = torch.from_numpy(lowercase__ ).to(lowercase__ ) self.timesteps += self.config.steps_offset def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : torch.FloatTensor ,lowercase__ : int ,lowercase__ : torch.FloatTensor ,lowercase__ : float = 0.0 ,lowercase__ : bool = False ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : bool = True ,): # 1. get previous step value (=t+1) __lowercase = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process __lowercase = self.alphas_cumprod[timestep] __lowercase = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) __lowercase = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": __lowercase = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 __lowercase = model_output elif self.config.prediction_type == "sample": __lowercase = model_output __lowercase = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": __lowercase = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output __lowercase = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( F"prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or" ''' `v_prediction`''' ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: __lowercase = pred_original_sample.clamp( -self.config.clip_sample_range ,self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __lowercase = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __lowercase = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=lowercase__ ,pred_original_sample=lowercase__ ) def __len__( self : List[Any] ): return self.config.num_train_timesteps
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
def _A ( A__ ): """simple docstring""" if number < 0: raise ValueError('''number must not be negative''' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ = { '''vocab_file''': {'''mobilebert-uncased''': '''https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt'''}, '''tokenizer_file''': { '''mobilebert-uncased''': '''https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json''' }, } lowerCAmelCase__ = {'''mobilebert-uncased''': 512} lowerCAmelCase__ = {} class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : int = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : List[str] = MobileBertTokenizer def __init__( self : Optional[Any] ,lowercase__ : Optional[Any]=None ,lowercase__ : Any=None ,lowercase__ : Dict=True ,lowercase__ : List[str]="[UNK]" ,lowercase__ : Dict="[SEP]" ,lowercase__ : Optional[Any]="[PAD]" ,lowercase__ : Dict="[CLS]" ,lowercase__ : str="[MASK]" ,lowercase__ : Optional[Any]=True ,lowercase__ : List[Any]=None ,**lowercase__ : Optional[Any] ,): super().__init__( lowercase__ ,tokenizer_file=lowercase__ ,do_lower_case=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,tokenize_chinese_chars=lowercase__ ,strip_accents=lowercase__ ,**lowercase__ ,) __lowercase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' ,lowercase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' ,lowercase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' ,lowercase__ ) != tokenize_chinese_chars ): __lowercase = getattr(lowercase__ ,normalizer_state.pop('''type''' ) ) __lowercase = do_lower_case __lowercase = strip_accents __lowercase = tokenize_chinese_chars __lowercase = normalizer_class(**lowercase__ ) __lowercase = do_lower_case def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Dict ,lowercase__ : Union[str, Any]=None ): __lowercase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : str ,lowercase__ : Optional[str] = None ): __lowercase = self._tokenizer.model.save(lowercase__ ,name=lowercase__ ) return tuple(lowercase__ )
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : str = ShapEPipeline SCREAMING_SNAKE_CASE : Optional[Any] = ['prompt'] SCREAMING_SNAKE_CASE : Tuple = ['prompt'] SCREAMING_SNAKE_CASE : int = [ 'num_images_per_prompt', 'num_inference_steps', 'generator', 'latents', 'guidance_scale', 'frame_size', 'output_type', 'return_dict', ] SCREAMING_SNAKE_CASE : Optional[int] = False @property def SCREAMING_SNAKE_CASE ( self : Tuple ): return 3_2 @property def SCREAMING_SNAKE_CASE ( self : str ): return 3_2 @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.time_input_dim * 4 @property def SCREAMING_SNAKE_CASE ( self : List[Any] ): return 8 @property def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def SCREAMING_SNAKE_CASE ( self : List[Any] ): torch.manual_seed(0 ) __lowercase = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=self.text_embedder_hidden_size ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=3_7 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_0_0_0 ,) return CLIPTextModelWithProjection(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): torch.manual_seed(0 ) __lowercase = { '''num_attention_heads''': 2, '''attention_head_dim''': 1_6, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 3_2, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } __lowercase = PriorTransformer(**lowercase__ ) return model @property def SCREAMING_SNAKE_CASE ( self : List[str] ): torch.manual_seed(0 ) __lowercase = { '''param_shapes''': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 1_2, '''background''': ( 0.1, 0.1, 0.1, ), } __lowercase = ShapERenderer(**lowercase__ ) return model def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.dummy_prior __lowercase = self.dummy_text_encoder __lowercase = self.dummy_tokenizer __lowercase = self.dummy_renderer __lowercase = HeunDiscreteScheduler( beta_schedule='''exp''' ,num_train_timesteps=1_0_2_4 ,prediction_type='''sample''' ,use_karras_sigmas=lowercase__ ,clip_sample=lowercase__ ,clip_sample_range=1.0 ,) __lowercase = { '''prior''': prior, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''renderer''': renderer, '''scheduler''': scheduler, } return components def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Tuple ,lowercase__ : Tuple=0 ): if str(lowercase__ ).startswith('''mps''' ): __lowercase = torch.manual_seed(lowercase__ ) else: __lowercase = torch.Generator(device=lowercase__ ).manual_seed(lowercase__ ) __lowercase = { '''prompt''': '''horse''', '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 3_2, '''output_type''': '''np''', } return inputs def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = '''cpu''' __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**lowercase__ ) __lowercase = pipe.to(lowercase__ ) pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = pipe(**self.get_dummy_inputs(lowercase__ ) ) __lowercase = output.images[0] __lowercase = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) __lowercase = np.array( [ 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self : str ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = torch_device == '''cpu''' __lowercase = True self._test_inference_batch_single_identical( batch_size=2 ,test_max_difference=lowercase__ ,relax_max_difference=lowercase__ ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**lowercase__ ) __lowercase = pipe.to(lowercase__ ) pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = 1 __lowercase = 2 __lowercase = self.get_dummy_inputs(lowercase__ ) for key in inputs.keys(): if key in self.batch_params: __lowercase = batch_size * [inputs[key]] __lowercase = pipe(**lowercase__ ,num_images_per_prompt=lowercase__ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : int ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_np_out.npy''' ) __lowercase = ShapEPipeline.from_pretrained('''openai/shap-e''' ) __lowercase = pipe.to(lowercase__ ) pipe.set_progress_bar_config(disable=lowercase__ ) __lowercase = torch.Generator(device=lowercase__ ).manual_seed(0 ) __lowercase = pipe( '''a shark''' ,generator=lowercase__ ,guidance_scale=1_5.0 ,num_inference_steps=6_4 ,frame_size=6_4 ,output_type='''np''' ,).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(lowercase__ ,lowercase__ )
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ ): """simple docstring""" __lowercase = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: __lowercase = 1024 __lowercase = 4096 __lowercase = 24 __lowercase = 16 __lowercase = [5, 11, 17, 23] __lowercase = [256, 512, 1024, 1024] __lowercase = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: __lowercase = 768 __lowercase = [1, 1, 1, 0.5] __lowercase = [256, 512, 768, 768] __lowercase = 150 __lowercase = 16 __lowercase = (1, 384, 384) __lowercase = False __lowercase = '''project''' if "ade" in checkpoint_url: __lowercase = True __lowercase = 768 __lowercase = [1, 1, 1, 0.5] __lowercase = 150 __lowercase = 16 __lowercase = '''huggingface/label-files''' __lowercase = '''ade20k-id2label.json''' __lowercase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type='''dataset''' ) ) , '''r''' ) ) __lowercase = {int(A__ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} __lowercase = [1, 150, 480, 480] return config, expected_shape def _A ( A__ ): """simple docstring""" __lowercase = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(A__ , A__ ) def _A ( A__ ): """simple docstring""" if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowercase = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: __lowercase = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: __lowercase = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: __lowercase = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: __lowercase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: __lowercase = name.replace('''proj''' , '''projection''' ) if "blocks" in name: __lowercase = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: __lowercase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: __lowercase = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: __lowercase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: __lowercase = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: __lowercase = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: __lowercase = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: __lowercase = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: __lowercase = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: __lowercase = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: __lowercase = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: __lowercase = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowercase = name.replace(F"refinenet{layer_idx}" , F"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: __lowercase = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: __lowercase = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: __lowercase = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: __lowercase = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: __lowercase = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: __lowercase = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowercase = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: __lowercase = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: __lowercase = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: __lowercase = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: __lowercase = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: __lowercase = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: __lowercase = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: __lowercase = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: __lowercase = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: __lowercase = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: __lowercase = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: __lowercase = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: __lowercase = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: __lowercase = name.replace('''..''' , '''.''' ) if "stem.conv" in name: __lowercase = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: __lowercase = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: __lowercase = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: __lowercase = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: __lowercase = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: __lowercase = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: __lowercase = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def _A ( A__ , A__ ): """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowercase = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.weight" ) __lowercase = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase = in_proj_weight[: config.hidden_size, :] __lowercase = in_proj_bias[: config.hidden_size] __lowercase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowercase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowercase = in_proj_weight[ -config.hidden_size :, : ] __lowercase = in_proj_bias[-config.hidden_size :] def _A ( ): """simple docstring""" __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase , __lowercase = get_dpt_config(A__ ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") __lowercase = torch.load(A__ , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowercase = state_dict.pop(A__ ) __lowercase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowercase = DPTForSemanticSegmentation(A__ ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowercase = 480 if '''ade''' in checkpoint_url else 384 __lowercase = DPTImageProcessor(size=A__ ) __lowercase = prepare_img() __lowercase = image_processor(A__ , return_tensors='''pt''' ) # forward pass __lowercase = model(**A__ ).logits if '''ade''' in checkpoint_url else model(**A__ ).predicted_depth if show_prediction: __lowercase = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=A__ , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(A__ ).mkdir(exist_ok=A__ ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(A__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(A__ ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=False, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) parser.add_argument( '''--show_prediction''', action='''store_true''', ) lowerCAmelCase__ = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = 'facebook/bart-large-mnli' SCREAMING_SNAKE_CASE : Tuple = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) SCREAMING_SNAKE_CASE : List[str] = 'text_classifier' SCREAMING_SNAKE_CASE : int = AutoTokenizer SCREAMING_SNAKE_CASE : Any = AutoModelForSequenceClassification SCREAMING_SNAKE_CASE : Optional[int] = ['text', ['text']] SCREAMING_SNAKE_CASE : List[Any] = ['text'] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): super().setup() __lowercase = self.model.config __lowercase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail''' ): __lowercase = int(lowercase__ ) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Any ,lowercase__ : Optional[int] ): __lowercase = labels return self.pre_processor( [text] * len(lowercase__ ) ,[F"This example is {label}" for label in labels] ,return_tensors='''pt''' ,padding='''max_length''' ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Any ): __lowercase = outputs.logits __lowercase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from __future__ import annotations def _A ( A__ , A__ , A__ , ): """simple docstring""" if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative in a semiconductor''' ) elif hole_conc < 0: raise ValueError('''Hole concentration cannot be negative in a semiconductor''' ) elif intrinsic_conc < 0: raise ValueError( '''Intrinsic concentration cannot be negative in a semiconductor''' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' from math import factorial def _A ( A__ , A__ ): """simple docstring""" if n < k or k < 0: raise ValueError('''Please enter positive integers for n and k where n >= k''' ) return factorial(A__ ) // (factorial(A__ ) * factorial(n - k )) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f'fifty-two card deck is: {combinations(52, 5)}\n', ) print( '''If a class of 40 students must be arranged into groups of''', f'4 for group projects, there are {combinations(40, 4)} ways', '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f'are {combinations(10, 3)} ways that first, second and', '''third place can be awarded.''', )
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-1''' lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-2''' lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-3''' lowerCAmelCase__ = '''CompVis/stable-diffusion-v1-4''' class lowercase_ (lowerCamelCase__ ): def __init__( self : Optional[int] ,lowercase__ : AutoencoderKL ,lowercase__ : CLIPTextModel ,lowercase__ : CLIPTokenizer ,lowercase__ : UNetaDConditionModel ,lowercase__ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] ,lowercase__ : StableDiffusionSafetyChecker ,lowercase__ : CLIPImageProcessor ,lowercase__ : bool = True ,): super()._init_() __lowercase = StableDiffusionPipeline.from_pretrained(lowercase__ ) __lowercase = StableDiffusionPipeline.from_pretrained(lowercase__ ) __lowercase = StableDiffusionPipeline.from_pretrained(lowercase__ ) __lowercase = StableDiffusionPipeline( vae=lowercase__ ,text_encoder=lowercase__ ,tokenizer=lowercase__ ,unet=lowercase__ ,scheduler=lowercase__ ,safety_checker=lowercase__ ,feature_extractor=lowercase__ ,requires_safety_checker=lowercase__ ,) self.register_modules(pipelinea=self.pipea ,pipelinea=self.pipea ,pipelinea=self.pipea ,pipelinea=self.pipea ) @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return {k: getattr(self ,lowercase__ ) for k in self.config.keys() if not k.startswith('''_''' )} def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Optional[Union[str, int]] = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __lowercase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.enable_attention_slicing(lowercase__ ) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : int ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Optional[Any] ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Optional[int] ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : Any ,): return self.pipea( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Union[str, List[str]] ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_1_2 ,lowercase__ : int = 5_0 ,lowercase__ : float = 7.5 ,lowercase__ : Optional[Union[str, List[str]]] = None ,lowercase__ : Optional[int] = 1 ,lowercase__ : float = 0.0 ,lowercase__ : Optional[torch.Generator] = None ,lowercase__ : Optional[torch.FloatTensor] = None ,lowercase__ : Optional[str] = "pil" ,lowercase__ : bool = True ,lowercase__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,lowercase__ : int = 1 ,**lowercase__ : str ,): __lowercase = '''cuda''' if torch.cuda.is_available() else '''cpu''' self.to(lowercase__ ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` must be divisible by 8 but are {height} and {width}." ) # Get first result from Stable Diffusion Checkpoint v1.1 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get first result from Stable Diffusion Checkpoint v1.2 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get first result from Stable Diffusion Checkpoint v1.3 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get first result from Stable Diffusion Checkpoint v1.4 __lowercase = self.textaimg_sda_a( prompt=lowercase__ ,height=lowercase__ ,width=lowercase__ ,num_inference_steps=lowercase__ ,guidance_scale=lowercase__ ,negative_prompt=lowercase__ ,num_images_per_prompt=lowercase__ ,eta=lowercase__ ,generator=lowercase__ ,latents=lowercase__ ,output_type=lowercase__ ,return_dict=lowercase__ ,callback=lowercase__ ,callback_steps=lowercase__ ,**lowercase__ ,) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''facebook/xmod-base''': '''https://huggingface.co/facebook/xmod-base/resolve/main/config.json''', '''facebook/xmod-large-prenorm''': '''https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json''', '''facebook/xmod-base-13-125k''': '''https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json''', '''facebook/xmod-base-30-125k''': '''https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json''', '''facebook/xmod-base-30-195k''': '''https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json''', '''facebook/xmod-base-60-125k''': '''https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json''', '''facebook/xmod-base-60-265k''': '''https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json''', '''facebook/xmod-base-75-125k''': '''https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json''', '''facebook/xmod-base-75-269k''': '''https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json''', } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = 'xmod' def __init__( self : Dict ,lowercase__ : str=3_0_5_2_2 ,lowercase__ : List[Any]=7_6_8 ,lowercase__ : List[Any]=1_2 ,lowercase__ : Optional[Any]=1_2 ,lowercase__ : Tuple=3_0_7_2 ,lowercase__ : List[str]="gelu" ,lowercase__ : Dict=0.1 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : List[Any]=5_1_2 ,lowercase__ : List[Any]=2 ,lowercase__ : int=0.0_2 ,lowercase__ : int=1e-1_2 ,lowercase__ : Union[str, Any]=1 ,lowercase__ : List[str]=0 ,lowercase__ : Dict=2 ,lowercase__ : List[str]="absolute" ,lowercase__ : Dict=True ,lowercase__ : Any=None ,lowercase__ : Optional[Any]=False ,lowercase__ : List[str]=2 ,lowercase__ : List[str]=False ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=True ,lowercase__ : Any=("en_XX",) ,lowercase__ : int=None ,**lowercase__ : str ,): super().__init__(pad_token_id=lowercase__ ,bos_token_id=lowercase__ ,eos_token_id=lowercase__ ,**lowercase__ ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = classifier_dropout __lowercase = pre_norm __lowercase = adapter_reduction_factor __lowercase = adapter_layer_norm __lowercase = adapter_reuse_layer_norm __lowercase = ln_before_adapter __lowercase = list(lowercase__ ) __lowercase = default_language class lowercase_ (lowerCamelCase__ ): """simple docstring""" @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): if self.task == "multiple-choice": __lowercase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __lowercase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' def _A ( A__ = 10 ): """simple docstring""" if not isinstance(A__ , A__ ) or n < 0: raise ValueError('''Invalid input''' ) __lowercase = 10**n __lowercase = 28433 * (pow(2 , 7830457 , A__ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f'{solution(10) = }')
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' from __future__ import annotations lowerCAmelCase__ = [] def _A ( A__ , A__ , A__ ): """simple docstring""" for i in range(len(A__ ) ): if board[row][i] == 1: return False for i in range(len(A__ ) ): if board[i][column] == 1: return False for i, j in zip(range(A__ , -1 , -1 ) , range(A__ , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(A__ , -1 , -1 ) , range(A__ , len(A__ ) ) ): if board[i][j] == 1: return False return True def _A ( A__ , A__ ): """simple docstring""" if row >= len(A__ ): solution.append(A__ ) printboard(A__ ) print() return True for i in range(len(A__ ) ): if is_safe(A__ , A__ , A__ ): __lowercase = 1 solve(A__ , row + 1 ) __lowercase = 0 return False def _A ( A__ ): """simple docstring""" for i in range(len(A__ ) ): for j in range(len(A__ ) ): if board[i][j] == 1: print('''Q''' , end=''' ''' ) else: print('''.''' , end=''' ''' ) print() # n=int(input("The no. of queens")) lowerCAmelCase__ = 8 lowerCAmelCase__ = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('''The total no. of solutions are :''', len(solution))
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
import numpy as np def _A ( A__ , A__ , A__ = 1e-12 , A__ = 100 , ): """simple docstring""" assert np.shape(A__ )[0] == np.shape(A__ )[1] # Ensure proper dimensionality. assert np.shape(A__ )[0] == np.shape(A__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(A__ ) == np.iscomplexobj(A__ ) __lowercase = np.iscomplexobj(A__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(A__ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. __lowercase = False __lowercase = 0 __lowercase = 0 __lowercase = 1e12 while not convergence: # Multiple matrix by the vector. __lowercase = np.dot(A__ , A__ ) # Normalize the resulting output vector. __lowercase = w / np.linalg.norm(A__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) __lowercase = vector.conj().T if is_complex else vector.T __lowercase = np.dot(A__ , np.dot(A__ , A__ ) ) # Check convergence. __lowercase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: __lowercase = True __lowercase = lambda_ if is_complex: __lowercase = np.real(lambda_ ) return lambda_, vector def _A ( ): """simple docstring""" __lowercase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) __lowercase = np.array([41, 4, 20] ) __lowercase = real_input_matrix.astype(np.complexaaa ) __lowercase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T __lowercase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": __lowercase = real_input_matrix __lowercase = real_vector elif problem_type == "complex": __lowercase = complex_input_matrix __lowercase = complex_vector # Our implementation. __lowercase , __lowercase = power_iteration(A__ , A__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). __lowercase , __lowercase = np.linalg.eigh(A__ ) # Last eigenvalue is the maximum one. __lowercase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. __lowercase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(A__ ) - np.abs(A__ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = '''▁''' lowerCAmelCase__ = { '''vocab_file''': '''vocab.json''', '''spm_file''': '''sentencepiece.bpe.model''', } lowerCAmelCase__ = { '''vocab_file''': { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/vocab.json''' ), }, '''spm_file''': { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/sentencepiece.bpe.model''' ) }, } lowerCAmelCase__ = { '''facebook/s2t-small-librispeech-asr''': 1024, } lowerCAmelCase__ = ['''pt''', '''fr''', '''ru''', '''nl''', '''ro''', '''it''', '''es''', '''de'''] lowerCAmelCase__ = {'''mustc''': MUSTC_LANGS} class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : int = MAX_MODEL_INPUT_SIZES SCREAMING_SNAKE_CASE : int = ['input_ids', 'attention_mask'] SCREAMING_SNAKE_CASE : List[int] = [] def __init__( self : Optional[int] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : Tuple="<s>" ,lowercase__ : List[str]="</s>" ,lowercase__ : str="<pad>" ,lowercase__ : Optional[Any]="<unk>" ,lowercase__ : Dict=False ,lowercase__ : List[Any]=False ,lowercase__ : int=None ,lowercase__ : Dict=None ,lowercase__ : Optional[Dict[str, Any]] = None ,**lowercase__ : Optional[Any] ,): __lowercase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,pad_token=lowercase__ ,do_upper_case=lowercase__ ,do_lower_case=lowercase__ ,tgt_lang=lowercase__ ,lang_codes=lowercase__ ,sp_model_kwargs=self.sp_model_kwargs ,**lowercase__ ,) __lowercase = do_upper_case __lowercase = do_lower_case __lowercase = load_json(lowercase__ ) __lowercase = {v: k for k, v in self.encoder.items()} __lowercase = spm_file __lowercase = load_spm(lowercase__ ,self.sp_model_kwargs ) if lang_codes is not None: __lowercase = lang_codes __lowercase = LANGUAGES[lang_codes] __lowercase = [F"<lang:{lang}>" for lang in self.langs] __lowercase = {lang: self.sp_model.PieceToId(F"<lang:{lang}>" ) for lang in self.langs} __lowercase = self.lang_tokens __lowercase = tgt_lang if tgt_lang is not None else self.langs[0] self.set_tgt_lang_special_tokens(self._tgt_lang ) else: __lowercase = {} @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return len(self.encoder ) @property def SCREAMING_SNAKE_CASE ( self : int ): return self._tgt_lang @tgt_lang.setter def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : str ): __lowercase = new_tgt_lang self.set_tgt_lang_special_tokens(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : str ): __lowercase = self.lang_code_to_id[tgt_lang] __lowercase = [lang_code_id] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : str ): return self.sp_model.encode(lowercase__ ,out_type=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ): return self.encoder.get(lowercase__ ,self.encoder[self.unk_token] ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : int ): return self.decoder.get(lowercase__ ,self.unk_token ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ): __lowercase = [] __lowercase = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: __lowercase = self.sp_model.decode(lowercase__ ) out_string += (decoded.upper() if self.do_upper_case else decoded) + token + " " __lowercase = [] else: current_sub_tokens.append(lowercase__ ) __lowercase = self.sp_model.decode(lowercase__ ) out_string += decoded.upper() if self.do_upper_case else decoded return out_string.strip() def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Any ,lowercase__ : Optional[int]=None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase__ ,token_ids_a=lowercase__ ,already_has_special_tokens=lowercase__ ) __lowercase = [1] * len(self.prefix_tokens ) __lowercase = [1] if token_ids_a is None: return prefix_ones + ([0] * len(lowercase__ )) + suffix_ones return prefix_ones + ([0] * len(lowercase__ )) + ([0] * len(lowercase__ )) + suffix_ones def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.encoder.copy() vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : Tuple ,lowercase__ : Dict ): __lowercase = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs''' ): __lowercase = {} __lowercase = load_spm(self.spm_file ,self.sp_model_kwargs ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): __lowercase = Path(lowercase__ ) assert save_dir.is_dir(), F"{save_directory} should be a directory" __lowercase = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) __lowercase = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder ,lowercase__ ) if os.path.abspath(self.spm_file ) != os.path.abspath(lowercase__ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file ,lowercase__ ) elif not os.path.isfile(self.spm_file ): with open(lowercase__ ,'''wb''' ) as fi: __lowercase = self.sp_model.serialized_model_proto() fi.write(lowercase__ ) return (str(lowercase__ ), str(lowercase__ )) def _A ( A__ , A__ ): """simple docstring""" __lowercase = sentencepiece.SentencePieceProcessor(**A__ ) spm.Load(str(A__ ) ) return spm def _A ( A__ ): """simple docstring""" with open(A__ , '''r''' ) as f: return json.load(A__ ) def _A ( A__ , A__ ): """simple docstring""" with open(A__ , '''w''' ) as f: json.dump(A__ , A__ , indent=2 )
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
def _A ( A__ , A__ ): """simple docstring""" __lowercase = '''''' for i in table: res += inp[i - 1] return res def _A ( A__ ): """simple docstring""" return data[1:] + data[0] def _A ( A__ , A__ ): """simple docstring""" __lowercase = '''''' for i in range(len(A__ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def _A ( A__ , A__ ): """simple docstring""" __lowercase = int('''0b''' + data[0] + data[-1] , 2 ) __lowercase = int('''0b''' + data[1:3] , 2 ) return bin(s[row][col] )[2:] def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = message[:4] __lowercase = message[4:] __lowercase = apply_table(A__ , A__ ) __lowercase = xor(A__ , A__ ) __lowercase = apply_sbox(A__ , temp[:4] ) # noqa: E741 __lowercase = apply_sbox(A__ , temp[4:] ) __lowercase = '''0''' * (2 - len(A__ )) + l # noqa: E741 __lowercase = '''0''' * (2 - len(A__ )) + r __lowercase = apply_table(l + r , A__ ) __lowercase = xor(A__ , A__ ) return temp + right if __name__ == "__main__": lowerCAmelCase__ = input('''Enter 10 bit key: ''') lowerCAmelCase__ = input('''Enter 8 bit message: ''') lowerCAmelCase__ = [6, 3, 7, 4, 8, 5, 10, 9] lowerCAmelCase__ = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] lowerCAmelCase__ = [2, 4, 3, 1] lowerCAmelCase__ = [2, 6, 3, 1, 4, 8, 5, 7] lowerCAmelCase__ = [4, 1, 3, 5, 7, 2, 8, 6] lowerCAmelCase__ = [4, 1, 2, 3, 2, 3, 4, 1] lowerCAmelCase__ = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] lowerCAmelCase__ = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation lowerCAmelCase__ = apply_table(key, paa_table) lowerCAmelCase__ = temp[:5] lowerCAmelCase__ = temp[5:] lowerCAmelCase__ = left_shift(left) lowerCAmelCase__ = left_shift(right) lowerCAmelCase__ = apply_table(left + right, pa_table) lowerCAmelCase__ = left_shift(left) lowerCAmelCase__ = left_shift(right) lowerCAmelCase__ = left_shift(left) lowerCAmelCase__ = left_shift(right) lowerCAmelCase__ = apply_table(left + right, pa_table) # encryption lowerCAmelCase__ = apply_table(message, IP) lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = temp[4:] + temp[:4] lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = apply_table(temp, IP_inv) print('''Cipher text is:''', CT) # decryption lowerCAmelCase__ = apply_table(CT, IP) lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = temp[4:] + temp[:4] lowerCAmelCase__ = function(expansion, sa, sa, keya, temp) lowerCAmelCase__ = apply_table(temp, IP_inv) print('''Plain text after decypting is:''', PT)
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def _A ( A__ , A__ , A__ , A__ = 100 , ): """simple docstring""" __lowercase = x_start __lowercase = fnc(A__ ) __lowercase = 0.0 for _ in range(A__ ): # Approximates curve as a sequence of linear lines and sums their length __lowercase = (x_end - x_start) / steps + xa __lowercase = fnc(A__ ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step __lowercase = xa __lowercase = fxa return length if __name__ == "__main__": def _A ( A__ ): """simple docstring""" return math.sin(10 * x ) print('''f(x) = sin(10 * x)''') print('''The length of the curve from x = -10 to x = 10 is:''') lowerCAmelCase__ = 10 while i <= 10_0000: print(f'With {i} steps: {line_length(f, -10, 10, i)}') i *= 10
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 1 # (0 is vertical, 1 is horizontal) def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) print('''Processing...''' ) __lowercase , __lowercase , __lowercase = update_image_and_anno(A__ , A__ , A__ ) for index, image in enumerate(A__ ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = paths[index].split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}" cva.imwrite(F"/{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Success {index+1}/{len(A__ )} with {file_name}" ) __lowercase = [] for anno in new_annos[index]: __lowercase = F"{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}" annos_list.append(A__ ) with open(F"/{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ = 1 ): """simple docstring""" __lowercase = [] __lowercase = [] __lowercase = [] for idx in range(len(A__ ) ): __lowercase = [] __lowercase = img_list[idx] path_list.append(A__ ) __lowercase = anno_list[idx] __lowercase = cva.imread(A__ ) if flip_type == 1: __lowercase = cva.flip(A__ , A__ ) for bbox in img_annos: __lowercase = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: __lowercase = cva.flip(A__ , A__ ) for bbox in img_annos: __lowercase = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(A__ ) new_imgs_list.append(A__ ) return new_imgs_list, new_annos_lists, path_list def _A ( A__ = 32 ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : str ): with open(lowercase__ ,encoding='''utf-8''' ) as input_file: __lowercase = re.compile(r'''(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)''' ) __lowercase = input_file.read() __lowercase = regexp.search(lowercase__ ) return match def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : str ): with open(lowercase__ ,encoding='''utf-8''' ) as input_file: __lowercase = re.compile(r'''#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()''' ,re.DOTALL ) __lowercase = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __lowercase = regexp.finditer(lowercase__ ) __lowercase = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = Path('''./datasets''' ) __lowercase = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowercase__ ) ): raise AssertionError(F"open(...) must use utf-8 encoding in {dataset}" ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = Path('''./datasets''' ) __lowercase = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_print_statements(str(lowercase__ ) ): raise AssertionError(F"print statement found in {dataset}. Use datasets.logger/logging instead." )
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed lowerCAmelCase__ = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(f'{bindir}/../../examples/pytorch/translation'): from run_translation import main # noqa set_seed(42) lowerCAmelCase__ = '''sshleifer/student_marian_en_ro_6_1''' lowerCAmelCase__ = '''sshleifer/tiny-mbart''' @require_torch class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : int=False ,lowercase__ : Optional[Any]=None ,lowercase__ : int=True ,lowercase__ : int=True ,lowercase__ : Optional[Any]=True ,lowercase__ : str=True ,): __lowercase = self.run_trainer( eval_steps=1 ,max_len=1_2 ,model_name=lowercase__ ,num_train_epochs=1 ,distributed=lowercase__ ,extra_args_str=lowercase__ ,predict_with_generate=lowercase__ ,do_train=lowercase__ ,do_eval=lowercase__ ,do_predict=lowercase__ ,) __lowercase = TrainerState.load_from_json(os.path.join(lowercase__ ,'''trainer_state.json''' ) ).log_history if not do_eval: return __lowercase = [log for log in logs if '''eval_loss''' in log.keys()] __lowercase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats __lowercase = eval_metrics[-1] assert isinstance(last_step_stats['''eval_bleu'''] ,lowercase__ ) assert not math.isnan(float(last_step_stats['''eval_loss'''] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def SCREAMING_SNAKE_CASE ( self : str ): self.run_seqaseq_quick() @require_torch_multi_gpu def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): self.run_seqaseq_quick(distributed=lowercase__ ) @require_torch_multi_gpu def SCREAMING_SNAKE_CASE ( self : Dict ): self.run_seqaseq_quick(distributed=lowercase__ ) @unittest.skip('''Requires an update of the env running those tests''' ) @require_torch_multi_gpu @require_fairscale def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): self.run_seqaseq_quick(distributed=lowercase__ ,extra_args_str='''--sharded_ddp simple''' ) @unittest.skip('''Requires an update of the env running those tests''' ) @require_torch_multi_gpu @require_fairscale def SCREAMING_SNAKE_CASE ( self : int ): self.run_seqaseq_quick(distributed=lowercase__ ,extra_args_str='''--sharded_ddp simple --fp16''' ) @unittest.skip('''Requires an update of the env running those tests''' ) @require_torch_multi_gpu @require_fairscale def SCREAMING_SNAKE_CASE ( self : Dict ): self.run_seqaseq_quick(distributed=lowercase__ ,extra_args_str='''--sharded_ddp zero_dp_2''' ,predict_with_generate=lowercase__ ) @unittest.skip('''Requires an update of the env running those tests''' ) @require_torch_multi_gpu @require_fairscale def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.run_seqaseq_quick( distributed=lowercase__ ,extra_args_str='''--sharded_ddp zero_dp_2 --fp16''' ,predict_with_generate=lowercase__ ) @require_apex @require_torch_gpu def SCREAMING_SNAKE_CASE ( self : Tuple ): # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=lowercase__ ,extra_args_str='''--fp16 --fp16_backend=apex''' ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=lowercase__ ,extra_args_str='''--fp16 --fp16_backend=apex''' ) @parameterized.expand(['''base''', '''low''', '''high''', '''mixed'''] ) @require_torch_multi_gpu def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Dict ): # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout __lowercase = { # test with the default log_level - should be info and thus log info once '''base''': {'''extra_args_str''': '''''', '''n_matches''': 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes '''low''': {'''extra_args_str''': '''--log_level debug --log_level_replica debug''', '''n_matches''': 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica '''high''': {'''extra_args_str''': '''--log_level error --log_level_replica debug''', '''n_matches''': 1}, # test with high log_level and log_level_replica - should be quiet on all processes '''mixed''': {'''extra_args_str''': '''--log_level error --log_level_replica error''', '''n_matches''': 0}, } __lowercase = experiments[experiment_id] __lowercase = {'''distributed''': True, '''predict_with_generate''': False, '''do_eval''': False, '''do_predict''': False} __lowercase = '''Running training''' with CaptureStderr() as cl: self.run_seqaseq_quick(**lowercase__ ,extra_args_str=data['''extra_args_str'''] ) __lowercase = len(re.findall(lowercase__ ,cl.err ) ) self.assertEqual(lowercase__ ,data['''n_matches'''] ) @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.run_trainer( eval_steps=2 ,max_len=1_2_8 ,model_name=lowercase__ ,learning_rate=3e-4 ,num_train_epochs=1_0 ,distributed=lowercase__ ,) # Check metrics __lowercase = TrainerState.load_from_json(os.path.join(lowercase__ ,'''trainer_state.json''' ) ).log_history __lowercase = [log for log in logs if '''eval_loss''' in log.keys()] __lowercase = eval_metrics[0] __lowercase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats['''eval_bleu'''] ,lowercase__ ) # test if do_predict saves generations and metrics __lowercase = os.listdir(lowercase__ ) __lowercase = {os.path.basename(lowercase__ ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): from transformers.training_args import OptimizerNames def train_and_return_metrics(lowercase__ : str ) -> Tuple[int, float]: __lowercase = '''--skip_memory_metrics 0''' __lowercase = self.run_trainer( max_len=1_2_8 ,model_name=lowercase__ ,learning_rate=3e-4 ,num_train_epochs=1 ,optim=lowercase__ ,distributed=lowercase__ ,extra_args_str=lowercase__ ,do_eval=lowercase__ ,do_predict=lowercase__ ,n_gpus_to_use=1 ,) # Check metrics __lowercase = TrainerState.load_from_json(Path(lowercase__ ,'''trainer_state.json''' ) ).log_history __lowercase = int(logs[0]['''train_mem_gpu_peaked_delta'''] / 2**2_0 ) __lowercase = int(logs[0]['''train_mem_gpu_alloc_delta'''] / 2**2_0 ) __lowercase = logs[0]['''train_loss'''] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss __lowercase , __lowercase , __lowercase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) __lowercase , __lowercase , __lowercase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) __lowercase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb __lowercase = gpu_peak_mem_orig + gpu_alloc_mem_orig __lowercase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb __lowercase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings __lowercase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( lowercase__ ,lowercase__ ,'''should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got''' F" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and" F" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB" ,) self.assertGreater( lowercase__ ,lowercase__ ,'''should use ~150MB less total gpu memory with BNB, compared to without it for this model but got''' F" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and" F" gpu_total_mem_bnb={gpu_total_mem_bnb}MB" ,) self.assertEqual( lowercase__ ,lowercase__ ,F"loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}" ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : int ,lowercase__ : str ,lowercase__ : int ,lowercase__ : float = 3e-3 ,lowercase__ : str = "adafactor" ,lowercase__ : bool = False ,lowercase__ : str = None ,lowercase__ : int = 0 ,lowercase__ : bool = True ,lowercase__ : bool = True ,lowercase__ : bool = True ,lowercase__ : bool = True ,lowercase__ : int = None ,): __lowercase = self.test_file_dir / '''../fixtures/tests_samples/wmt_en_ro''' __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n --model_name_or_path {model_name}\n --train_file {data_dir}/train.json\n --validation_file {data_dir}/val.json\n --test_file {data_dir}/test.json\n --output_dir {output_dir}\n --overwrite_output_dir\n --max_train_samples 8\n --max_source_length {max_len}\n --max_target_length {max_len}\n --do_train\n --num_train_epochs {str(lowercase__ )}\n --per_device_train_batch_size 4\n --learning_rate {learning_rate}\n --warmup_steps 8\n --logging_steps 0\n --logging_strategy no\n --save_steps {str(lowercase__ )}\n --group_by_length\n --label_smoothing_factor 0.1\n --target_lang ro_RO\n --source_lang en_XX\n ".split() __lowercase = F"\n --do_eval\n --per_device_eval_batch_size 4\n --max_eval_samples 8\n --val_max_target_length {max_len}\n --evaluation_strategy steps\n --eval_steps {str(lowercase__ )}\n ".split() __lowercase = ''' --do_predict '''.split() __lowercase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"--optim {optim}".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: __lowercase = get_gpu_count() __lowercase = get_torch_dist_unique_port() __lowercase = F"\n -m torch.distributed.run\n --nproc_per_node={n_gpus_to_use}\n --master_port={master_port}\n {self.examples_dir_str}/pytorch/translation/run_translation.py\n ".split() __lowercase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowercase__ ,env=self.get_env() ) else: __lowercase = ['''run_translation.py'''] + args with patch.object(lowercase__ ,'''argv''' ,lowercase__ ): main() return output_dir
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' from math import pi, sqrt def _A ( A__ ): """simple docstring""" if num <= 0: raise ValueError('''math domain error''' ) if num > 171.5: raise OverflowError('''math range error''' ) elif num - int(A__ ) not in (0, 0.5): raise NotImplementedError('''num must be an integer or a half-integer''' ) elif num == 0.5: return sqrt(A__ ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def _A ( ): """simple docstring""" assert gamma(0.5 ) == sqrt(A__ ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() lowerCAmelCase__ = 1.0 while num: lowerCAmelCase__ = float(input('''Gamma of: ''')) print(f'gamma({num}) = {gamma(num)}') print('''\nEnter 0 to exit...''')
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''google/mobilenet_v1_1.0_224''': '''https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json''', '''google/mobilenet_v1_0.75_192''': '''https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = 'mobilenet_v1' def __init__( self : Tuple ,lowercase__ : Tuple=3 ,lowercase__ : List[Any]=2_2_4 ,lowercase__ : str=1.0 ,lowercase__ : List[str]=8 ,lowercase__ : Tuple="relu6" ,lowercase__ : Optional[Any]=True ,lowercase__ : int=0.9_9_9 ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : List[str]=0.0_0_1 ,**lowercase__ : Union[str, Any] ,): super().__init__(**lowercase__ ) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''' ) __lowercase = num_channels __lowercase = image_size __lowercase = depth_multiplier __lowercase = min_depth __lowercase = hidden_act __lowercase = tf_padding __lowercase = classifier_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE ( self : Dict ): return OrderedDict([('''pixel_values''', {0: '''batch'''})] ) @property def SCREAMING_SNAKE_CASE ( self : str ): if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})] ) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})] ) @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return 1e-4
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def _A ( A__ , A__ ): """simple docstring""" __lowercase = old_name if "patch_embed" in old_name: __lowercase , __lowercase , __lowercase = old_name.split('''.''' ) if layer == "0": __lowercase = old_name.replace('''0''' , '''convolution1''' ) elif layer == "1": __lowercase = old_name.replace('''1''' , '''batchnorm_before''' ) elif layer == "3": __lowercase = old_name.replace('''3''' , '''convolution2''' ) else: __lowercase = old_name.replace('''4''' , '''batchnorm_after''' ) if "network" in old_name and re.search(R'''\d\.\d''' , A__ ): __lowercase = R'''\b\d{2}\b''' if bool(re.search(A__ , A__ ) ): __lowercase = re.search(R'''\d\.\d\d.''' , A__ ).group() else: __lowercase = re.search(R'''\d\.\d.''' , A__ ).group() if int(match[0] ) < 6: __lowercase = old_name.replace(A__ , '''''' ) __lowercase = trimmed_name.replace('''network''' , match[0] + '''.meta4D_layers.blocks.''' + match[2:-1] ) __lowercase = '''intermediate_stages.''' + trimmed_name else: __lowercase = old_name.replace(A__ , '''''' ) if int(match[2] ) < num_meta4D_last_stage: __lowercase = trimmed_name.replace('''network''' , '''meta4D_layers.blocks.''' + match[2] ) else: __lowercase = str(int(match[2] ) - num_meta4D_last_stage ) __lowercase = trimmed_name.replace('''network''' , '''meta3D_layers.blocks.''' + layer_index ) if "norm1" in old_name: __lowercase = trimmed_name.replace('''norm1''' , '''layernorm1''' ) elif "norm2" in old_name: __lowercase = trimmed_name.replace('''norm2''' , '''layernorm2''' ) elif "fc1" in old_name: __lowercase = trimmed_name.replace('''fc1''' , '''linear_in''' ) elif "fc2" in old_name: __lowercase = trimmed_name.replace('''fc2''' , '''linear_out''' ) __lowercase = '''last_stage.''' + trimmed_name elif "network" in old_name and re.search(R'''.\d.''' , A__ ): __lowercase = old_name.replace('''network''' , '''intermediate_stages''' ) if "fc" in new_name: __lowercase = new_name.replace('''fc''' , '''convolution''' ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): __lowercase = new_name.replace('''norm1''' , '''batchnorm_before''' ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): __lowercase = new_name.replace('''norm2''' , '''batchnorm_after''' ) if "proj" in new_name: __lowercase = new_name.replace('''proj''' , '''projection''' ) if "dist_head" in new_name: __lowercase = new_name.replace('''dist_head''' , '''distillation_classifier''' ) elif "head" in new_name: __lowercase = new_name.replace('''head''' , '''classifier''' ) elif "patch_embed" in new_name: __lowercase = '''efficientformer.''' + new_name elif new_name == "norm.weight" or new_name == "norm.bias": __lowercase = new_name.replace('''norm''' , '''layernorm''' ) __lowercase = '''efficientformer.''' + new_name else: __lowercase = '''efficientformer.encoder.''' + new_name return new_name def _A ( A__ , A__ ): """simple docstring""" for key in checkpoint.copy().keys(): __lowercase = checkpoint.pop(A__ ) __lowercase = val return checkpoint def _A ( ): """simple docstring""" __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = Image.open(requests.get(A__ , stream=A__ ).raw ) return image def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = torch.load(A__ , map_location='''cpu''' )['''model'''] __lowercase = EfficientFormerConfig.from_json_file(A__ ) __lowercase = EfficientFormerForImageClassificationWithTeacher(A__ ) __lowercase = '''_'''.join(checkpoint_path.split('''/''' )[-1].split('''.''' )[0].split('''_''' )[:-1] ) __lowercase = config.depths[-1] - config.num_metaad_blocks + 1 __lowercase = convert_torch_checkpoint(A__ , A__ ) model.load_state_dict(A__ ) model.eval() __lowercase = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } # prepare image __lowercase = prepare_img() __lowercase = 256 __lowercase = 224 __lowercase = EfficientFormerImageProcessor( size={'''shortest_edge''': image_size} , crop_size={'''height''': crop_size, '''width''': crop_size} , resample=pillow_resamplings['''bicubic'''] , ) __lowercase = processor(images=A__ , return_tensors='''pt''' ).pixel_values # original processing pipeline __lowercase = Compose( [ Resize(A__ , interpolation=pillow_resamplings['''bicubic'''] ), CenterCrop(A__ ), ToTensor(), Normalize(A__ , A__ ), ] ) __lowercase = image_transforms(A__ ).unsqueeze(0 ) assert torch.allclose(A__ , A__ ) __lowercase = model(A__ ) __lowercase = outputs.logits __lowercase = (1, 1000) if "l1" in model_name: __lowercase = torch.Tensor( [-0.1_3_1_2, 0.4_3_5_3, -1.0_4_9_9, -0.5_1_2_4, 0.4_1_8_3, -0.6_7_9_3, -1.3_7_7_7, -0.0_8_9_3, -0.7_3_5_8, -2.4_3_2_8] ) assert torch.allclose(logits[0, :10] , A__ , atol=1e-3 ) assert logits.shape == expected_shape elif "l3" in model_name: __lowercase = torch.Tensor( [-1.3_1_5_0, -1.5_4_5_6, -1.2_5_5_6, -0.8_4_9_6, -0.7_1_2_7, -0.7_8_9_7, -0.9_7_2_8, -0.3_0_5_2, 0.3_7_5_1, -0.3_1_2_7] ) assert torch.allclose(logits[0, :10] , A__ , atol=1e-3 ) assert logits.shape == expected_shape elif "l7" in model_name: __lowercase = torch.Tensor( [-1.0_2_8_3, -1.4_1_3_1, -0.5_6_4_4, -1.3_1_1_5, -0.5_7_8_5, -1.2_0_4_9, -0.7_5_2_8, 0.1_9_9_2, -0.3_8_2_2, -0.0_8_7_8] ) assert logits.shape == expected_shape else: raise ValueError( F"Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7" ) # Save Checkpoints Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) print(F"Checkpoint successfuly converted. Model saved at {pytorch_dump_path}" ) processor.save_pretrained(A__ ) print(F"Processor successfuly saved at {pytorch_dump_path}" ) if push_to_hub: print('''Pushing model to the hub...''' ) model.push_to_hub( repo_id=F"Bearnardd/{pytorch_dump_path}" , commit_message='''Add model''' , use_temp_dir=A__ , ) processor.push_to_hub( repo_id=F"Bearnardd/{pytorch_dump_path}" , commit_message='''Add image processor''' , use_temp_dir=A__ , ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--pytorch_model_path''', default=None, type=str, required=True, help='''Path to EfficientFormer pytorch checkpoint.''', ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The json file for EfficientFormer model config.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') parser.add_argument( '''--no-push_to_hub''', dest='''push_to_hub''', action='''store_false''', help='''Do not push model and image processor to the hub''', ) parser.set_defaults(push_to_hub=True) lowerCAmelCase__ = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent lowerCAmelCase__ = {'''UserAgent''': UserAgent().random} def _A ( A__ ): """simple docstring""" __lowercase = script.contents[0] __lowercase = json.loads(data[data.find('''{"config"''' ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : Optional[int] ): __lowercase = F"https://www.instagram.com/{username}/" __lowercase = self.get_json() def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = requests.get(self.url ,headers=lowercase__ ).text __lowercase = BeautifulSoup(lowercase__ ,'''html.parser''' ).find_all('''script''' ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self : Optional[int] ): return F"{self.__class__.__name__}('{self.username}')" def __str__( self : List[str] ): return F"{self.fullname} ({self.username}) is {self.biography}" @property def SCREAMING_SNAKE_CASE ( self : Tuple ): return self.user_data["username"] @property def SCREAMING_SNAKE_CASE ( self : int ): return self.user_data["full_name"] @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.user_data["biography"] @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.user_data["business_email"] @property def SCREAMING_SNAKE_CASE ( self : List[str] ): return self.user_data["external_url"] @property def SCREAMING_SNAKE_CASE ( self : List[str] ): return self.user_data["edge_followed_by"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Dict ): return self.user_data["edge_follow"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return self.user_data["edge_owner_to_timeline_media"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return self.user_data["profile_pic_url_hd"] @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return self.user_data["is_verified"] @property def SCREAMING_SNAKE_CASE ( self : Tuple ): return self.user_data["is_private"] def _A ( A__ = "github" ): """simple docstring""" import os if os.environ.get('''CI''' ): return # test failing on GitHub Actions __lowercase = InstagramUser(A__ ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , A__ ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 150 assert instagram_user.number_of_followers > 120000 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith('''https://instagram.''' ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = InstagramUser('''github''') print(instagram_user) print(f'{instagram_user.number_of_posts = }') print(f'{instagram_user.number_of_followers = }') print(f'{instagram_user.number_of_followings = }') print(f'{instagram_user.email = }') print(f'{instagram_user.website = }') print(f'{instagram_user.profile_picture_url = }') print(f'{instagram_user.is_verified = }') print(f'{instagram_user.is_private = }')
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def _A ( A__ , A__ , A__ , A__=5 ): """simple docstring""" assert masked_input.count('''<mask>''' ) == 1 __lowercase = torch.tensor(tokenizer.encode(A__ , add_special_tokens=A__ ) ).unsqueeze(0 ) # Batch size 1 __lowercase = model(A__ )[0] # The last hidden-state is the first element of the output tuple __lowercase = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() __lowercase = logits[0, masked_index, :] __lowercase = logits.softmax(dim=0 ) __lowercase , __lowercase = prob.topk(k=A__ , dim=0 ) __lowercase = ''' '''.join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(A__ ) )] ) __lowercase = tokenizer.mask_token __lowercase = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(''' ''' ) ): __lowercase = predicted_token_bpe.replace('''\u2581''' , ''' ''' ) if " {0}".format(A__ ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(''' {0}'''.format(A__ ) , A__ ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(A__ , A__ ), values[index].item(), predicted_token, ) ) return topk_filled_outputs lowerCAmelCase__ = CamembertTokenizer.from_pretrained('''camembert-base''') lowerCAmelCase__ = CamembertForMaskedLM.from_pretrained('''camembert-base''') model.eval() lowerCAmelCase__ = '''Le camembert est <mask> :)''' print(fill_mask(masked_input, model, tokenizer, topk=3))
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ = { '''vocab_file''': { '''gpt2''': '''https://huggingface.co/gpt2/resolve/main/vocab.json''', '''gpt2-medium''': '''https://huggingface.co/gpt2-medium/resolve/main/vocab.json''', '''gpt2-large''': '''https://huggingface.co/gpt2-large/resolve/main/vocab.json''', '''gpt2-xl''': '''https://huggingface.co/gpt2-xl/resolve/main/vocab.json''', '''distilgpt2''': '''https://huggingface.co/distilgpt2/resolve/main/vocab.json''', }, '''merges_file''': { '''gpt2''': '''https://huggingface.co/gpt2/resolve/main/merges.txt''', '''gpt2-medium''': '''https://huggingface.co/gpt2-medium/resolve/main/merges.txt''', '''gpt2-large''': '''https://huggingface.co/gpt2-large/resolve/main/merges.txt''', '''gpt2-xl''': '''https://huggingface.co/gpt2-xl/resolve/main/merges.txt''', '''distilgpt2''': '''https://huggingface.co/distilgpt2/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''gpt2''': '''https://huggingface.co/gpt2/resolve/main/tokenizer.json''', '''gpt2-medium''': '''https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json''', '''gpt2-large''': '''https://huggingface.co/gpt2-large/resolve/main/tokenizer.json''', '''gpt2-xl''': '''https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json''', '''distilgpt2''': '''https://huggingface.co/distilgpt2/resolve/main/tokenizer.json''', }, } lowerCAmelCase__ = { '''gpt2''': 1024, '''gpt2-medium''': 1024, '''gpt2-large''': 1024, '''gpt2-xl''': 1024, '''distilgpt2''': 1024, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Tuple = ['input_ids', 'attention_mask'] SCREAMING_SNAKE_CASE : Optional[int] = GPTaTokenizer def __init__( self : int ,lowercase__ : List[str]=None ,lowercase__ : List[Any]=None ,lowercase__ : Any=None ,lowercase__ : List[str]="<|endoftext|>" ,lowercase__ : List[Any]="<|endoftext|>" ,lowercase__ : int="<|endoftext|>" ,lowercase__ : Optional[Any]=False ,**lowercase__ : Dict ,): super().__init__( lowercase__ ,lowercase__ ,tokenizer_file=lowercase__ ,unk_token=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,add_prefix_space=lowercase__ ,**lowercase__ ,) __lowercase = kwargs.pop('''add_bos_token''' ,lowercase__ ) __lowercase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' ,lowercase__ ) != add_prefix_space: __lowercase = getattr(lowercase__ ,pre_tok_state.pop('''type''' ) ) __lowercase = add_prefix_space __lowercase = pre_tok_class(**lowercase__ ) __lowercase = add_prefix_space def SCREAMING_SNAKE_CASE ( self : Tuple ,*lowercase__ : Optional[int] ,**lowercase__ : Dict ): __lowercase = kwargs.get('''is_split_into_words''' ,lowercase__ ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ,*lowercase__ : Tuple ,**lowercase__ : Union[str, Any] ): __lowercase = kwargs.get('''is_split_into_words''' ,lowercase__ ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : str ,lowercase__ : Optional[str] = None ): __lowercase = self._tokenizer.model.save(lowercase__ ,name=lowercase__ ) return tuple(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : "Conversation" ): __lowercase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowercase__ ,add_special_tokens=lowercase__ ) + [self.eos_token_id] ) if len(lowercase__ ) > self.model_max_length: __lowercase = input_ids[-self.model_max_length :] return input_ids
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = 'vivit' def __init__( self : Union[str, Any] ,lowercase__ : Optional[Any]=2_2_4 ,lowercase__ : Any=3_2 ,lowercase__ : Optional[Any]=[2, 1_6, 1_6] ,lowercase__ : Any=3 ,lowercase__ : str=7_6_8 ,lowercase__ : Any=1_2 ,lowercase__ : int=1_2 ,lowercase__ : Any=3_0_7_2 ,lowercase__ : Optional[Any]="gelu_fast" ,lowercase__ : str=0.0 ,lowercase__ : Dict=0.0 ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : int=1e-0_6 ,lowercase__ : int=True ,**lowercase__ : Dict ,): __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = num_frames __lowercase = tubelet_size __lowercase = num_channels __lowercase = qkv_bias super().__init__(**lowercase__ )
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''hidden_sizes''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_attention_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Any ,lowercase__ : Optional[int]=1_3 ,lowercase__ : Union[str, Any]=6_4 ,lowercase__ : List[str]=3 ,lowercase__ : Any=3 ,lowercase__ : Any=2 ,lowercase__ : Optional[int]=1 ,lowercase__ : Tuple=1_6 ,lowercase__ : List[str]=[1_2_8, 2_5_6, 3_8_4] ,lowercase__ : List[str]=[4, 6, 8] ,lowercase__ : Tuple=[2, 3, 4] ,lowercase__ : str=[1_6, 1_6, 1_6] ,lowercase__ : Optional[int]=0 ,lowercase__ : List[Any]=[2, 2, 2] ,lowercase__ : List[str]=[2, 2, 2] ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : List[str]=True ,lowercase__ : Optional[Any]=True ,lowercase__ : Any=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = kernel_size __lowercase = stride __lowercase = padding __lowercase = hidden_sizes __lowercase = num_attention_heads __lowercase = depths __lowercase = key_dim __lowercase = drop_path_rate __lowercase = patch_size __lowercase = attention_ratio __lowercase = mlp_ratio __lowercase = initializer_range __lowercase = [ ['''Subsample''', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['''Subsample''', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = initializer_range def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : List[Any] ): return LevitConfig( image_size=self.image_size ,num_channels=self.num_channels ,kernel_size=self.kernel_size ,stride=self.stride ,padding=self.padding ,patch_size=self.patch_size ,hidden_sizes=self.hidden_sizes ,num_attention_heads=self.num_attention_heads ,depths=self.depths ,key_dim=self.key_dim ,drop_path_rate=self.drop_path_rate ,mlp_ratio=self.mlp_ratio ,attention_ratio=self.attention_ratio ,initializer_range=self.initializer_range ,down_ops=self.down_ops ,) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ,lowercase__ : Any ): __lowercase = LevitModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for _ in range(4 ): __lowercase = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) __lowercase = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) ,) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : str ): __lowercase = self.num_labels __lowercase = LevitForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': LevitModel, 'image-classification': (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Tuple = False def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = LevitModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : Any ): return @unittest.skip(reason='''Levit does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass @unittest.skip(reason='''Levit does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass @unittest.skip(reason='''Levit does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): def check_hidden_states_output(lowercase__ : str ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowercase__ ) ,lowercase__ ) __lowercase = (self.model_tester.image_size, self.model_tester.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for _ in range(4 ): __lowercase = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) __lowercase = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[ height * width, self.model_tester.hidden_sizes[0], ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : Tuple ,lowercase__ : Tuple=False ): __lowercase = super()._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): if not self.model_tester.is_training: return __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowercase__ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : int ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowercase = False __lowercase = True for model_class in self.all_model_classes: if model_class in get_values(lowercase__ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue __lowercase = model_class(lowercase__ ) model.gradient_checkpointing_enable() model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) __lowercase = model(**lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowercase__ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): __lowercase = problem_type['''title'''] __lowercase = problem_type['''num_labels'''] __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = self._prepare_for_class(lowercase__ ,lowercase__ ,return_labels=lowercase__ ) if problem_type["num_labels"] > 1: __lowercase = inputs['''labels'''].unsqueeze(1 ).repeat(1 ,problem_type['''num_labels'''] ) __lowercase = inputs['''labels'''].to(problem_type['''dtype'''] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowercase__ ) as warning_list: __lowercase = model(**lowercase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = LevitModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Any ): return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([1.0_4_4_8, -0.3_7_4_5, -1.8_3_1_7] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowercase_ (lowerCamelCase__ ): SCREAMING_SNAKE_CASE : Any = ['image_processor', 'tokenizer'] SCREAMING_SNAKE_CASE : Optional[Any] = 'BridgeTowerImageProcessor' SCREAMING_SNAKE_CASE : Dict = ('RobertaTokenizer', 'RobertaTokenizerFast') def __init__( self : List[str] ,lowercase__ : int ,lowercase__ : Optional[Any] ): super().__init__(lowercase__ ,lowercase__ ) def __call__( self : Any ,lowercase__ : Union[str, Any] ,lowercase__ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None ,lowercase__ : bool = True ,lowercase__ : Union[bool, str, PaddingStrategy] = False ,lowercase__ : Union[bool, str, TruncationStrategy] = None ,lowercase__ : Optional[int] = None ,lowercase__ : int = 0 ,lowercase__ : Optional[int] = None ,lowercase__ : Optional[bool] = None ,lowercase__ : Optional[bool] = None ,lowercase__ : bool = False ,lowercase__ : bool = False ,lowercase__ : bool = False ,lowercase__ : bool = False ,lowercase__ : bool = True ,lowercase__ : Optional[Union[str, TensorType]] = None ,**lowercase__ : int ,): __lowercase = self.tokenizer( text=lowercase__ ,add_special_tokens=lowercase__ ,padding=lowercase__ ,truncation=lowercase__ ,max_length=lowercase__ ,stride=lowercase__ ,pad_to_multiple_of=lowercase__ ,return_token_type_ids=lowercase__ ,return_attention_mask=lowercase__ ,return_overflowing_tokens=lowercase__ ,return_special_tokens_mask=lowercase__ ,return_offsets_mapping=lowercase__ ,return_length=lowercase__ ,verbose=lowercase__ ,return_tensors=lowercase__ ,**lowercase__ ,) # add pixel_values + pixel_mask __lowercase = self.image_processor( lowercase__ ,return_tensors=lowercase__ ,do_normalize=lowercase__ ,do_center_crop=lowercase__ ,**lowercase__ ) encoding.update(lowercase__ ) return encoding def SCREAMING_SNAKE_CASE ( self : int ,*lowercase__ : List[Any] ,**lowercase__ : List[str] ): return self.tokenizer.batch_decode(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,*lowercase__ : Any ,**lowercase__ : List[str] ): return self.tokenizer.decode(*lowercase__ ,**lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.tokenizer.model_input_names __lowercase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' import string import numpy def _A ( A__ , A__ ): """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , A__ ) class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : int = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) SCREAMING_SNAKE_CASE : Dict = numpy.vectorize(lambda lowerCamelCase__ : x % 3_6 ) SCREAMING_SNAKE_CASE : Optional[Any] = numpy.vectorize(lowerCamelCase__ ) def __init__( self : Any ,lowercase__ : numpy.ndarray ): __lowercase = self.modulus(lowercase__ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key __lowercase = encrypt_key.shape[0] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : str ): return self.key_string.index(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : int ): return self.key_string[round(lowercase__ )] def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: __lowercase = det % len(self.key_string ) __lowercase = len(self.key_string ) if greatest_common_divisor(lowercase__ ,len(self.key_string ) ) != 1: __lowercase = ( F"determinant modular {req_l} of encryption key({det}) " F"is not co prime w.r.t {req_l}.\nTry another key." ) raise ValueError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : str ): __lowercase = [char for char in text.upper() if char in self.key_string] __lowercase = chars[-1] while len(lowercase__ ) % self.break_key != 0: chars.append(lowercase__ ) return "".join(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): __lowercase = self.process_text(text.upper() ) __lowercase = '''''' for i in range(0 ,len(lowercase__ ) - self.break_key + 1 ,self.break_key ): __lowercase = text[i : i + self.break_key] __lowercase = [self.replace_letters(lowercase__ ) for char in batch] __lowercase = numpy.array([vec] ).T __lowercase = self.modulus(self.encrypt_key.dot(lowercase__ ) ).T.tolist()[ 0 ] __lowercase = ''''''.join( self.replace_digits(lowercase__ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: __lowercase = det % len(self.key_string ) __lowercase = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: __lowercase = i break __lowercase = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : str ): __lowercase = self.make_decrypt_key() __lowercase = self.process_text(text.upper() ) __lowercase = '''''' for i in range(0 ,len(lowercase__ ) - self.break_key + 1 ,self.break_key ): __lowercase = text[i : i + self.break_key] __lowercase = [self.replace_letters(lowercase__ ) for char in batch] __lowercase = numpy.array([vec] ).T __lowercase = self.modulus(decrypt_key.dot(lowercase__ ) ).T.tolist()[0] __lowercase = ''''''.join( self.replace_digits(lowercase__ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def _A ( ): """simple docstring""" __lowercase = int(input('''Enter the order of the encryption key: ''' ) ) __lowercase = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(A__ ): __lowercase = [int(A__ ) for x in input().split()] hill_matrix.append(A__ ) __lowercase = HillCipher(numpy.array(A__ ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) __lowercase = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": __lowercase = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(A__ ) ) elif option == "2": __lowercase = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(A__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCAmelCase__ = HfArgumentParser(InitializationArguments) lowerCAmelCase__ = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowerCAmelCase__ = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowerCAmelCase__ = { '''vocab_size''': len(tokenizer), '''scale_attn_by_inverse_layer_idx''': True, '''reorder_and_upcast_attn''': True, } # Load model config (GPT-2 large in this case) lowerCAmelCase__ = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowerCAmelCase__ = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' import re from filelock import FileLock try: import nltk lowerCAmelCase__ = True except (ImportError, ModuleNotFoundError): lowerCAmelCase__ = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def _A ( A__ ): """simple docstring""" re.sub('''<n>''' , '''''' , A__ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(A__ ) )
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
import argparse import math import os import torch from neural_compressor.utils.pytorch import load from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, StableDiffusionPipeline, UNetaDConditionModel def _A ( ): """simple docstring""" __lowercase = argparse.ArgumentParser() parser.add_argument( '''-m''' , '''--pretrained_model_name_or_path''' , type=A__ , default=A__ , required=A__ , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , ) parser.add_argument( '''-c''' , '''--caption''' , type=A__ , default='''robotic cat with wings''' , help='''Text used to generate images.''' , ) parser.add_argument( '''-n''' , '''--images_num''' , type=A__ , default=4 , help='''How much images to generate.''' , ) parser.add_argument( '''-s''' , '''--seed''' , type=A__ , default=42 , help='''Seed for random process.''' , ) parser.add_argument( '''-ci''' , '''--cuda_id''' , type=A__ , default=0 , help='''cuda_id.''' , ) __lowercase = parser.parse_args() return args def _A ( A__ , A__ , A__ ): """simple docstring""" if not len(A__ ) == rows * cols: raise ValueError('''The specified number of rows and columns are not correct.''' ) __lowercase , __lowercase = imgs[0].size __lowercase = Image.new('''RGB''' , size=(cols * w, rows * h) ) __lowercase , __lowercase = grid.size for i, img in enumerate(A__ ): grid.paste(A__ , box=(i % cols * w, i // cols * h) ) return grid def _A ( A__ , A__="robotic cat with wings" , A__=7.5 , A__=50 , A__=1 , A__=42 , ): """simple docstring""" __lowercase = torch.Generator(pipeline.device ).manual_seed(A__ ) __lowercase = pipeline( A__ , guidance_scale=A__ , num_inference_steps=A__ , generator=A__ , num_images_per_prompt=A__ , ).images __lowercase = int(math.sqrt(A__ ) ) __lowercase = image_grid(A__ , rows=_rows , cols=num_images_per_prompt // _rows ) return grid, images lowerCAmelCase__ = parse_args() # Load models and create wrapper for stable diffusion lowerCAmelCase__ = CLIPTokenizer.from_pretrained(args.pretrained_model_name_or_path, subfolder='''tokenizer''') lowerCAmelCase__ = CLIPTextModel.from_pretrained(args.pretrained_model_name_or_path, subfolder='''text_encoder''') lowerCAmelCase__ = AutoencoderKL.from_pretrained(args.pretrained_model_name_or_path, subfolder='''vae''') lowerCAmelCase__ = UNetaDConditionModel.from_pretrained(args.pretrained_model_name_or_path, subfolder='''unet''') lowerCAmelCase__ = StableDiffusionPipeline.from_pretrained( args.pretrained_model_name_or_path, text_encoder=text_encoder, vae=vae, unet=unet, tokenizer=tokenizer ) lowerCAmelCase__ = lambda images, clip_input: (images, False) if os.path.exists(os.path.join(args.pretrained_model_name_or_path, '''best_model.pt''')): lowerCAmelCase__ = load(args.pretrained_model_name_or_path, model=unet) unet.eval() setattr(pipeline, '''unet''', unet) else: lowerCAmelCase__ = unet.to(torch.device('''cuda''', args.cuda_id)) lowerCAmelCase__ = pipeline.to(unet.device) lowerCAmelCase__ , lowerCAmelCase__ = generate_images(pipeline, prompt=args.caption, num_images_per_prompt=args.images_num, seed=args.seed) grid.save(os.path.join(args.pretrained_model_name_or_path, '''{}.png'''.format('''_'''.join(args.caption.split())))) lowerCAmelCase__ = os.path.join(args.pretrained_model_name_or_path, '''_'''.join(args.caption.split())) os.makedirs(dirname, exist_ok=True) for idx, image in enumerate(images): image.save(os.path.join(dirname, '''{}.png'''.format(idx + 1)))
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class lowercase_ : """simple docstring""" def __init__( self : List[Any] ,lowercase__ : Any ,lowercase__ : Tuple=2 ,lowercase__ : Union[str, Any]=True ,lowercase__ : Any=False ,lowercase__ : List[str]=1_0 ,lowercase__ : Dict=3 ,lowercase__ : Optional[Any]=3_2 * 4 ,lowercase__ : Union[str, Any]=3_2 * 6 ,lowercase__ : Union[str, Any]=4 ,lowercase__ : List[Any]=3_2 ,): __lowercase = parent __lowercase = batch_size __lowercase = is_training __lowercase = use_auxiliary_loss __lowercase = num_queries __lowercase = num_channels __lowercase = min_size __lowercase = max_size __lowercase = num_labels __lowercase = mask_feature_size def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( lowercase__ ) __lowercase = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=lowercase__ ) __lowercase = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=lowercase__ ) > 0.5 ).float() __lowercase = (torch.rand((self.batch_size, self.num_labels) ,device=lowercase__ ) > 0.5).long() __lowercase = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] ,) ,decoder_config=DetrConfig( decoder_ffn_dim=1_2_8 ,num_queries=self.num_queries ,decoder_attention_heads=2 ,d_model=self.mask_feature_size ,) ,mask_feature_size=self.mask_feature_size ,fpn_feature_size=self.mask_feature_size ,num_channels=self.num_channels ,num_labels=self.num_labels ,) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = self.prepare_config_and_inputs() __lowercase = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : Any ): __lowercase = output.encoder_hidden_states __lowercase = output.pixel_decoder_hidden_states __lowercase = output.transformer_decoder_hidden_states self.parent.assertTrue(len(lowercase__ ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowercase__ ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowercase__ ) ,config.decoder_config.decoder_layers ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Tuple ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : Tuple=False ): with torch.no_grad(): __lowercase = MaskFormerModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(pixel_values=lowercase__ ,pixel_mask=lowercase__ ) __lowercase = model(lowercase__ ,output_hidden_states=lowercase__ ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.mask_feature_size) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ): __lowercase = MaskFormerForInstanceSegmentation(config=lowercase__ ) model.to(lowercase__ ) model.eval() def comm_check_on_output(lowercase__ : str ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): __lowercase = model(pixel_values=lowercase__ ,pixel_mask=lowercase__ ) __lowercase = model(lowercase__ ) comm_check_on_output(lowercase__ ) __lowercase = model( pixel_values=lowercase__ ,pixel_mask=lowercase__ ,mask_labels=lowercase__ ,class_labels=lowercase__ ) comm_check_on_output(lowercase__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': MaskFormerModel, 'image-segmentation': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : str = False SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = MaskFormerModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(lowercase__ ,**lowercase__ ,output_hidden_states=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*lowercase__ ) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass @unittest.skip(reason='''MaskFormer is not a generative model''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass @unittest.skip(reason='''MaskFormer does not use token embeddings''' ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : int ): for model_name in ["facebook/maskformer-swin-small-coco"]: __lowercase = MaskFormerModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = (self.model_tester.min_size,) * 2 __lowercase = { '''pixel_values''': torch.randn((2, 3, *size) ,device=lowercase__ ), '''mask_labels''': torch.randn((2, 1_0, *size) ,device=lowercase__ ), '''class_labels''': torch.zeros(2 ,1_0 ,device=lowercase__ ).long(), } __lowercase = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(lowercase__ ) __lowercase = model(**lowercase__ ) self.assertTrue(outputs.loss is not None ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(lowercase__ ,**lowercase__ ,output_hidden_states=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ).to(lowercase__ ) __lowercase = model(**lowercase__ ,output_attentions=lowercase__ ) self.assertTrue(outputs.attentions is not None ) def SCREAMING_SNAKE_CASE ( self : int ): if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss __lowercase = self.all_model_classes[1] __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs() __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = model(lowercase__ ,mask_labels=lowercase__ ,class_labels=lowercase__ ).loss loss.backward() def SCREAMING_SNAKE_CASE ( self : str ): # only MaskFormerForInstanceSegmentation has the loss __lowercase = self.all_model_classes[1] __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs() __lowercase = True __lowercase = True __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.train() __lowercase = model(lowercase__ ,mask_labels=lowercase__ ,class_labels=lowercase__ ) __lowercase = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() __lowercase = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't __lowercase = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() __lowercase = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=lowercase__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) lowerCAmelCase__ = 1e-4 def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : List[Any] ): return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''' ) if is_vision_available() else None ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''' ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) __lowercase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(lowercase__ ,(1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): __lowercase = model(**lowercase__ ) __lowercase = torch.tensor( [[-0.0_4_8_2, 0.9_2_2_8, 0.4_9_5_1], [-0.2_5_4_7, 0.8_0_1_7, 0.8_5_2_7], [-0.0_0_6_9, 0.3_3_8_5, -0.0_0_8_9]] ).to(lowercase__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) __lowercase = torch.tensor( [[-0.8_4_2_2, -0.8_4_3_4, -0.9_7_1_8], [-1.0_1_4_4, -0.5_5_6_5, -0.4_1_9_5], [-1.0_0_3_8, -0.4_4_8_4, -0.1_9_6_1]] ).to(lowercase__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) __lowercase = torch.tensor( [[0.2_8_5_2, -0.0_1_5_9, 0.9_7_3_5], [0.6_2_5_4, 0.1_8_5_8, 0.8_5_2_9], [-0.0_6_8_0, -0.4_1_1_6, 1.8_4_1_3]] ).to(lowercase__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(lowercase__ ) .eval() ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) __lowercase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(lowercase__ ,(1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): __lowercase = model(**lowercase__ ) # masks_queries_logits __lowercase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ,) __lowercase = [ [-1.3_7_3_7_1_2_4, -1.7_7_2_4_9_3_7, -1.9_3_6_4_2_3_3], [-1.5_9_7_7_2_8_1, -1.9_8_6_7_9_3_9, -2.1_5_2_3_6_9_5], [-1.5_7_9_5_3_9_8, -1.9_2_6_9_8_3_2, -2.0_9_3_9_4_2], ] __lowercase = torch.tensor(lowercase__ ).to(lowercase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) # class_queries_logits __lowercase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape ,(1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) __lowercase = torch.tensor( [ [1.6_5_1_2e0_0, -5.2_5_7_2e0_0, -3.3_5_1_9e0_0], [3.6_1_6_9e-0_2, -5.9_0_2_5e0_0, -2.9_3_1_3e0_0], [1.0_7_6_6e-0_4, -7.7_6_3_0e0_0, -5.1_2_6_3e0_0], ] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''' ) .to(lowercase__ ) .eval() ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) __lowercase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(lowercase__ ,(1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): __lowercase = model(**lowercase__ ) # masks_queries_logits __lowercase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ,) __lowercase = [[-0.9_0_4_6, -2.6_3_6_6, -4.6_0_6_2], [-3.4_1_7_9, -5.7_8_9_0, -8.8_0_5_7], [-4.9_1_7_9, -7.6_5_6_0, -1_0.7_7_1_1]] __lowercase = torch.tensor(lowercase__ ).to(lowercase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) # class_queries_logits __lowercase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape ,(1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) __lowercase = torch.tensor( [[4.7_1_8_8, -3.2_5_8_5, -2.8_8_5_7], [6.6_8_7_1, -2.9_1_8_1, -1.2_4_8_7], [7.2_4_4_9, -2.2_7_6_4, -2.1_8_7_4]] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,lowercase__ ,atol=lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(lowercase__ ) .eval() ) __lowercase = self.default_image_processor __lowercase = image_processor( [np.zeros((3, 8_0_0, 1_3_3_3) ), np.zeros((3, 8_0_0, 1_3_3_3) )] ,segmentation_maps=[np.zeros((3_8_4, 3_8_4) ).astype(np.floataa ), np.zeros((3_8_4, 3_8_4) ).astype(np.floataa )] ,return_tensors='''pt''' ,) __lowercase = inputs['''pixel_values'''].to(lowercase__ ) __lowercase = [el.to(lowercase__ ) for el in inputs['''mask_labels''']] __lowercase = [el.to(lowercase__ ) for el in inputs['''class_labels''']] with torch.no_grad(): __lowercase = model(**lowercase__ ) self.assertTrue(outputs.loss is not None )
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
def _A ( A__ = 3 , A__ = 7 , A__ = 1000000 ): """simple docstring""" __lowercase = 0 __lowercase = 1 for current_denominator in range(1 , limit + 1 ): __lowercase = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: __lowercase = current_numerator __lowercase = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=100_0000))
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase : Union[str, Any] = logging.get_logger(__name__) _lowercase : Optional[Any] = { "microsoft/wavlm-base": "https://huggingface.co/microsoft/wavlm-base/resolve/main/config.json", # See all WavLM models at https://huggingface.co/models?filter=wavlm } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 'wavlm' def __init__( self : List[str], lowerCamelCase : int=32, lowerCamelCase : List[str]=768, lowerCamelCase : List[str]=12, lowerCamelCase : Union[str, Any]=12, lowerCamelCase : Dict=3072, lowerCamelCase : int="gelu", lowerCamelCase : Union[str, Any]=0.1, lowerCamelCase : Union[str, Any]=0.1, lowerCamelCase : List[Any]=0.1, lowerCamelCase : Any=0.0, lowerCamelCase : Optional[Any]=0.1, lowerCamelCase : Union[str, Any]=0.1, lowerCamelCase : Dict=0.02, lowerCamelCase : Union[str, Any]=1E-5, lowerCamelCase : Tuple="group", lowerCamelCase : List[str]="gelu", lowerCamelCase : Any=(512, 512, 512, 512, 512, 512, 512), lowerCamelCase : Tuple=(5, 2, 2, 2, 2, 2, 2), lowerCamelCase : List[Any]=(10, 3, 3, 3, 3, 2, 2), lowerCamelCase : int=False, lowerCamelCase : str=128, lowerCamelCase : Optional[Any]=16, lowerCamelCase : str=320, lowerCamelCase : List[Any]=800, lowerCamelCase : List[Any]=False, lowerCamelCase : Optional[int]=True, lowerCamelCase : Optional[int]=0.05, lowerCamelCase : int=10, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : Any=0.0, lowerCamelCase : List[str]=10, lowerCamelCase : Optional[int]=320, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : Dict=0.1, lowerCamelCase : List[Any]=100, lowerCamelCase : List[Any]=256, lowerCamelCase : Union[str, Any]=256, lowerCamelCase : str=0.1, lowerCamelCase : int="mean", lowerCamelCase : Union[str, Any]=False, lowerCamelCase : int=False, lowerCamelCase : int=256, lowerCamelCase : Tuple=(512, 512, 512, 512, 1500), lowerCamelCase : Tuple=(5, 3, 3, 1, 1), lowerCamelCase : Optional[int]=(1, 2, 3, 1, 1), lowerCamelCase : str=512, lowerCamelCase : int=80, lowerCamelCase : Any=0, lowerCamelCase : int=1, lowerCamelCase : int=2, lowerCamelCase : Union[str, Any]=False, lowerCamelCase : Optional[Any]=3, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : List[str]=3, lowerCamelCase : Tuple=None, **lowerCamelCase : int, )-> Dict: super().__init__(**lowerCamelCase, pad_token_id=lowerCamelCase, bos_token_id=lowerCamelCase, eos_token_id=lowerCamelCase ) lowerCamelCase__ : str =hidden_size lowerCamelCase__ : str =feat_extract_norm lowerCamelCase__ : Tuple =feat_extract_activation lowerCamelCase__ : Optional[int] =list(lowerCamelCase ) lowerCamelCase__ : Tuple =list(lowerCamelCase ) lowerCamelCase__ : Dict =list(lowerCamelCase ) lowerCamelCase__ : Optional[int] =conv_bias lowerCamelCase__ : Optional[int] =num_buckets lowerCamelCase__ : str =max_bucket_distance lowerCamelCase__ : Optional[Any] =num_conv_pos_embeddings lowerCamelCase__ : Union[str, Any] =num_conv_pos_embedding_groups lowerCamelCase__ : Any =len(self.conv_dim ) lowerCamelCase__ : int =num_hidden_layers lowerCamelCase__ : Optional[int] =intermediate_size lowerCamelCase__ : Union[str, Any] =hidden_act lowerCamelCase__ : str =num_attention_heads lowerCamelCase__ : Tuple =hidden_dropout lowerCamelCase__ : Optional[Any] =attention_dropout lowerCamelCase__ : Dict =activation_dropout lowerCamelCase__ : Tuple =feat_proj_dropout lowerCamelCase__ : List[str] =final_dropout lowerCamelCase__ : List[str] =layerdrop lowerCamelCase__ : Dict =layer_norm_eps lowerCamelCase__ : int =initializer_range lowerCamelCase__ : Optional[int] =num_ctc_classes lowerCamelCase__ : Tuple =vocab_size lowerCamelCase__ : int =do_stable_layer_norm lowerCamelCase__ : Any =use_weighted_layer_sum lowerCamelCase__ : Tuple =classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowerCamelCase__ : List[str] =apply_spec_augment lowerCamelCase__ : List[str] =mask_time_prob lowerCamelCase__ : Union[str, Any] =mask_time_length lowerCamelCase__ : Any =mask_time_min_masks lowerCamelCase__ : Optional[Any] =mask_feature_prob lowerCamelCase__ : Union[str, Any] =mask_feature_length # parameters for pretraining with codevector quantized representations lowerCamelCase__ : int =num_codevectors_per_group lowerCamelCase__ : Tuple =num_codevector_groups lowerCamelCase__ : Union[str, Any] =contrastive_logits_temperature lowerCamelCase__ : Any =num_negatives lowerCamelCase__ : Dict =codevector_dim lowerCamelCase__ : Dict =proj_codevector_dim lowerCamelCase__ : Any =diversity_loss_weight # ctc loss lowerCamelCase__ : Optional[Any] =ctc_loss_reduction lowerCamelCase__ : int =ctc_zero_infinity # adapter lowerCamelCase__ : Union[str, Any] =add_adapter lowerCamelCase__ : Tuple =adapter_kernel_size lowerCamelCase__ : Optional[Any] =adapter_stride lowerCamelCase__ : str =num_adapter_layers lowerCamelCase__ : Union[str, Any] =output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowerCamelCase__ : List[Any] =classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowerCamelCase__ : Optional[Any] =list(lowerCamelCase ) lowerCamelCase__ : int =list(lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =list(lowerCamelCase ) lowerCamelCase__ : List[str] =xvector_output_dim @property def snake_case ( self : Optional[int] )-> List[Any]: return functools.reduce(operator.mul, self.conv_stride, 1 )
625
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" if isinstance(__lowerCamelCase , collections.abc.Iterable ): return x return (x, x) @require_flax class __SCREAMING_SNAKE_CASE : '''simple docstring''' def snake_case ( self : Dict, lowerCamelCase : List[str], lowerCamelCase : Any )-> Union[str, Any]: pass def snake_case ( self : List[str] )-> List[str]: pass def snake_case ( self : Optional[Any] )-> str: pass def snake_case ( self : Union[str, Any], lowerCamelCase : np.ndarray, lowerCamelCase : np.ndarray, lowerCamelCase : float )-> Dict: lowerCamelCase__ : Union[str, Any] =np.abs((a - b) ).max() self.assertLessEqual(lowerCamelCase, lowerCamelCase, F'''Difference between torch and flax is {diff} (>= {tol}).''' ) def snake_case ( self : Dict, lowerCamelCase : Tuple, lowerCamelCase : Any, lowerCamelCase : List[str], lowerCamelCase : Dict, lowerCamelCase : Any=None, **lowerCamelCase : str )-> int: lowerCamelCase__ : List[str] =VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Tuple =FlaxVisionTextDualEncoderModel(lowerCamelCase ) lowerCamelCase__ : Dict =model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase ) self.assertEqual(output['''text_embeds'''].shape, (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output['''image_embeds'''].shape, (pixel_values.shape[0], config.projection_dim) ) def snake_case ( self : Any, lowerCamelCase : int, lowerCamelCase : Optional[Any], lowerCamelCase : List[str], lowerCamelCase : Union[str, Any], lowerCamelCase : str=None, **lowerCamelCase : List[Any] )-> int: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] =self.get_vision_text_model(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Union[str, Any] ={'''vision_model''': vision_model, '''text_model''': text_model} lowerCamelCase__ : Tuple =FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase ) lowerCamelCase__ : Optional[int] =model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase ) self.assertEqual(output['''text_embeds'''].shape, (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['''image_embeds'''].shape, (pixel_values.shape[0], model.config.projection_dim) ) def snake_case ( self : Any, lowerCamelCase : Dict, lowerCamelCase : Dict, lowerCamelCase : Tuple, lowerCamelCase : Union[str, Any], lowerCamelCase : Dict=None, **lowerCamelCase : int )-> List[str]: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] =self.get_vision_text_model(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Optional[int] ={'''vision_model''': vision_model, '''text_model''': text_model} lowerCamelCase__ : int =FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase ) lowerCamelCase__ : List[Any] =model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase ) lowerCamelCase__ : int =output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase ) lowerCamelCase__ : Dict =FlaxVisionTextDualEncoderModel.from_pretrained(lowerCamelCase ) lowerCamelCase__ : Optional[int] =model(input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase ) lowerCamelCase__ : List[str] =after_output[0] lowerCamelCase__ : Any =np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase, 1E-3 ) def snake_case ( self : Optional[Any], lowerCamelCase : Dict, lowerCamelCase : str, lowerCamelCase : str, lowerCamelCase : str, lowerCamelCase : List[Any]=None, **lowerCamelCase : List[Any] )-> Tuple: lowerCamelCase__ , lowerCamelCase__ : Optional[int] =self.get_vision_text_model(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Any ={'''vision_model''': vision_model, '''text_model''': text_model} lowerCamelCase__ : int =FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase ) lowerCamelCase__ : List[str] =model( input_ids=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, output_attentions=lowerCamelCase ) lowerCamelCase__ : int =output.vision_model_output.attentions self.assertEqual(len(lowerCamelCase ), vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase__ : Tuple =to_atuple(vision_model.config.image_size ) lowerCamelCase__ : Optional[Any] =to_atuple(vision_model.config.patch_size ) lowerCamelCase__ : Union[str, Any] =(image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) lowerCamelCase__ : int =num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:], (vision_config.num_attention_heads, seq_len, seq_len) ) lowerCamelCase__ : List[Any] =output.text_model_output.attentions self.assertEqual(len(lowerCamelCase ), text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:], (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]), ) def snake_case ( self : Tuple, lowerCamelCase : Optional[int], lowerCamelCase : Any, lowerCamelCase : Union[str, Any] )-> Any: pt_model.to(lowerCamelCase ) pt_model.eval() # prepare inputs lowerCamelCase__ : Any =inputs_dict lowerCamelCase__ : Any ={k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): lowerCamelCase__ : List[str] =pt_model(**lowerCamelCase ).to_tuple() lowerCamelCase__ : Optional[Any] =fx_model(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ), len(lowerCamelCase ), '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(fx_outputs[:4], pt_outputs[:4] ): self.assert_almost_equals(lowerCamelCase, pt_output.numpy(), 4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowerCamelCase ) lowerCamelCase__ : Optional[int] =FlaxVisionTextDualEncoderModel.from_pretrained(lowerCamelCase, from_pt=lowerCamelCase ) lowerCamelCase__ : List[Any] =fx_model_loaded(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ), len(lowerCamelCase ), '''Output lengths differ between Flax and PyTorch''' ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4], pt_outputs[:4] ): self.assert_almost_equals(lowerCamelCase, pt_output.numpy(), 4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowerCamelCase ) lowerCamelCase__ : str =VisionTextDualEncoderModel.from_pretrained(lowerCamelCase, from_flax=lowerCamelCase ) pt_model_loaded.to(lowerCamelCase ) pt_model_loaded.eval() with torch.no_grad(): lowerCamelCase__ : List[Any] =pt_model_loaded(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ), len(lowerCamelCase ), '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output_loaded in zip(fx_outputs[:4], pt_outputs_loaded[:4] ): self.assert_almost_equals(lowerCamelCase, pt_output_loaded.numpy(), 4E-2 ) def snake_case ( self : str, lowerCamelCase : Union[str, Any], lowerCamelCase : Optional[Any], lowerCamelCase : str )-> List[Any]: lowerCamelCase__ : Any =VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : List[Any] =VisionTextDualEncoderModel(lowerCamelCase ) lowerCamelCase__ : List[str] =FlaxVisionTextDualEncoderModel(lowerCamelCase ) lowerCamelCase__ : str =convert_pytorch_state_dict_to_flax(pt_model.state_dict(), lowerCamelCase ) lowerCamelCase__ : Tuple =fx_state self.check_pt_flax_equivalence(lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : List[Any], lowerCamelCase : Tuple, lowerCamelCase : Union[str, Any], lowerCamelCase : Union[str, Any] )-> Optional[int]: lowerCamelCase__ : Dict =VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Tuple =VisionTextDualEncoderModel(lowerCamelCase ) lowerCamelCase__ : List[str] =FlaxVisionTextDualEncoderModel(lowerCamelCase ) lowerCamelCase__ : Tuple =load_flax_weights_in_pytorch_model(lowerCamelCase, fx_model.params ) self.check_pt_flax_equivalence(lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : Optional[int] )-> Union[str, Any]: lowerCamelCase__ : Any =self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowerCamelCase ) def snake_case ( self : Tuple )-> int: lowerCamelCase__ : int =self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowerCamelCase ) def snake_case ( self : Tuple )-> Any: lowerCamelCase__ : Tuple =self.prepare_config_and_inputs() self.check_save_load(**lowerCamelCase ) def snake_case ( self : str )-> Any: lowerCamelCase__ : str =self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowerCamelCase ) @is_pt_flax_cross_test def snake_case ( self : Tuple )-> List[Any]: lowerCamelCase__ : Union[str, Any] =self.prepare_config_and_inputs() lowerCamelCase__ : Union[str, Any] =config_inputs_dict.pop('''vision_config''' ) lowerCamelCase__ : Optional[Any] =config_inputs_dict.pop('''text_config''' ) lowerCamelCase__ : Tuple =config_inputs_dict self.check_equivalence_pt_to_flax(lowerCamelCase, lowerCamelCase, lowerCamelCase ) self.check_equivalence_flax_to_pt(lowerCamelCase, lowerCamelCase, lowerCamelCase ) @slow def snake_case ( self : Optional[Any] )-> Tuple: lowerCamelCase__ , lowerCamelCase__ : Dict =self.get_pretrained_model_and_inputs() lowerCamelCase__ : Optional[int] =model_a(**lowerCamelCase ) lowerCamelCase__ : List[str] =outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowerCamelCase ) lowerCamelCase__ : int =FlaxVisionTextDualEncoderModel.from_pretrained(lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =model_a(**lowerCamelCase ) lowerCamelCase__ : List[Any] =after_outputs[0] lowerCamelCase__ : Any =np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase, 1E-5 ) @require_flax class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' def snake_case ( self : Optional[int] )-> Optional[Any]: lowerCamelCase__ : str =FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-vit''', '''hf-internal-testing/tiny-bert''', vision_from_pt=lowerCamelCase, text_from_pt=lowerCamelCase, ) lowerCamelCase__ : Union[str, Any] =13 lowerCamelCase__ : List[str] =floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) lowerCamelCase__ : List[str] =ids_tensor([batch_size, 4], model.config.text_config.vocab_size ) lowerCamelCase__ : Optional[int] =random_attention_mask([batch_size, 4] ) lowerCamelCase__ : Any ={'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def snake_case ( self : str, lowerCamelCase : str, lowerCamelCase : int )-> int: lowerCamelCase__ : str =FlaxViTModel(lowerCamelCase ) lowerCamelCase__ : Any =FlaxBertModel(lowerCamelCase ) return vision_model, text_model def snake_case ( self : int )-> Optional[int]: lowerCamelCase__ : Any =FlaxViTModelTester(self ) lowerCamelCase__ : Union[str, Any] =FlaxBertModelTester(self ) lowerCamelCase__ : Any =vit_model_tester.prepare_config_and_inputs() lowerCamelCase__ : Optional[Any] =bert_model_tester.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ : Any =vision_config_and_inputs lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Tuple =text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' def snake_case ( self : Optional[int] )-> Optional[int]: lowerCamelCase__ : Union[str, Any] =FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-clip''', '''hf-internal-testing/tiny-bert''', vision_from_pt=lowerCamelCase, text_from_pt=lowerCamelCase, ) lowerCamelCase__ : Union[str, Any] =13 lowerCamelCase__ : Optional[Any] =floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) lowerCamelCase__ : Union[str, Any] =ids_tensor([batch_size, 4], model.config.text_config.vocab_size ) lowerCamelCase__ : str =random_attention_mask([batch_size, 4] ) lowerCamelCase__ : Optional[int] ={'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def snake_case ( self : List[str], lowerCamelCase : Any, lowerCamelCase : Dict )-> Dict: lowerCamelCase__ : str =FlaxCLIPVisionModel(lowerCamelCase ) lowerCamelCase__ : Optional[Any] =FlaxBertModel(lowerCamelCase ) return vision_model, text_model def snake_case ( self : Optional[int] )-> Optional[Any]: lowerCamelCase__ : List[Any] =FlaxCLIPVisionModelTester(self ) lowerCamelCase__ : List[Any] =FlaxBertModelTester(self ) lowerCamelCase__ : Any =clip_model_tester.prepare_config_and_inputs() lowerCamelCase__ : Optional[int] =bert_model_tester.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ : List[Any] =vision_config_and_inputs lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] =text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def snake_case ( self : Tuple )-> Optional[Any]: lowerCamelCase__ : Any =FlaxVisionTextDualEncoderModel.from_pretrained('''clip-italian/clip-italian''', logit_scale_init_value=1.0 ) lowerCamelCase__ : List[Any] =VisionTextDualEncoderProcessor.from_pretrained('''clip-italian/clip-italian''' ) lowerCamelCase__ : Optional[int] =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowerCamelCase__ : Dict =processor( text=['''una foto di un gatto''', '''una foto di un cane'''], images=lowerCamelCase, padding=lowerCamelCase, return_tensors='''np''' ) lowerCamelCase__ : List[Any] =model(**lowerCamelCase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape, (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape, (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]), ) lowerCamelCase__ : Any =np.array([[1.2_284_727, 0.3_104_122]] ) self.assertTrue(np.allclose(outputs.logits_per_image, lowerCamelCase, atol=1E-3 ) )
625
1
"""simple docstring""" def snake_case__ ( __lowerCamelCase : int ): """simple docstring""" if isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError('''\'float\' object cannot be interpreted as an integer''' ) if isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError('''\'str\' object cannot be interpreted as an integer''' ) if num == 0: return "0b0" lowerCamelCase__ : Union[str, Any] =False if num < 0: lowerCamelCase__ : List[Any] =True lowerCamelCase__ : Union[str, Any] =-num lowerCamelCase__ : list[int] =[] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(__lowerCamelCase ) for e in binary ) return "0b" + "".join(str(__lowerCamelCase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
625
"""simple docstring""" def snake_case__ ( __lowerCamelCase : list , __lowerCamelCase : list , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int ): """simple docstring""" if index == number_of_items: return 0 lowerCamelCase__ : Optional[int] =0 lowerCamelCase__ : Union[str, Any] =0 lowerCamelCase__ : List[str] =knapsack(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index + 1 ) if weights[index] <= max_weight: lowerCamelCase__ : Dict =values[index] + knapsack( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , max_weight - weights[index] , index + 1 ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
625
1
"""simple docstring""" from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase : Tuple = logging.get_logger(__name__) _lowercase : List[str] = { "google/efficientnet-b7": "https://huggingface.co/google/efficientnet-b7/resolve/main/config.json", } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 'efficientnet' def __init__( self : str, lowerCamelCase : int = 3, lowerCamelCase : int = 600, lowerCamelCase : float = 2.0, lowerCamelCase : float = 3.1, lowerCamelCase : int = 8, lowerCamelCase : List[int] = [3, 3, 5, 3, 5, 5, 3], lowerCamelCase : List[int] = [32, 16, 24, 40, 80, 112, 192], lowerCamelCase : List[int] = [16, 24, 40, 80, 112, 192, 320], lowerCamelCase : List[int] = [], lowerCamelCase : List[int] = [1, 2, 2, 2, 1, 2, 1], lowerCamelCase : List[int] = [1, 2, 2, 3, 3, 4, 1], lowerCamelCase : List[int] = [1, 6, 6, 6, 6, 6, 6], lowerCamelCase : float = 0.25, lowerCamelCase : str = "swish", lowerCamelCase : int = 2560, lowerCamelCase : str = "mean", lowerCamelCase : float = 0.02, lowerCamelCase : float = 0.001, lowerCamelCase : float = 0.99, lowerCamelCase : float = 0.5, lowerCamelCase : float = 0.2, **lowerCamelCase : Tuple, )-> Optional[int]: super().__init__(**lowerCamelCase ) lowerCamelCase__ : int =num_channels lowerCamelCase__ : int =image_size lowerCamelCase__ : int =width_coefficient lowerCamelCase__ : List[Any] =depth_coefficient lowerCamelCase__ : Optional[int] =depth_divisor lowerCamelCase__ : Optional[int] =kernel_sizes lowerCamelCase__ : str =in_channels lowerCamelCase__ : Optional[Any] =out_channels lowerCamelCase__ : Optional[int] =depthwise_padding lowerCamelCase__ : int =strides lowerCamelCase__ : str =num_block_repeats lowerCamelCase__ : List[Any] =expand_ratios lowerCamelCase__ : Dict =squeeze_expansion_ratio lowerCamelCase__ : Dict =hidden_act lowerCamelCase__ : int =hidden_dim lowerCamelCase__ : str =pooling_type lowerCamelCase__ : str =initializer_range lowerCamelCase__ : List[str] =batch_norm_eps lowerCamelCase__ : List[Any] =batch_norm_momentum lowerCamelCase__ : Tuple =dropout_rate lowerCamelCase__ : Tuple =drop_connect_rate lowerCamelCase__ : Union[str, Any] =sum(lowerCamelCase ) * 4 class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = version.parse('1.11' ) @property def snake_case ( self : Any )-> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def snake_case ( self : Optional[Any] )-> float: return 1E-5
625
"""simple docstring""" _lowercase : Optional[Any] = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", "dataclasses": "dataclasses", "datasets": "datasets!=2.5.0", "decord": "decord==0.6.0", "deepspeed": "deepspeed>=0.9.3", "diffusers": "diffusers", "dill": "dill<0.3.5", "evaluate": "evaluate>=0.2.0", "fairscale": "fairscale>0.3", "faiss-cpu": "faiss-cpu", "fastapi": "fastapi", "filelock": "filelock", "flax": "flax>=0.4.1,<=0.7.0", "ftfy": "ftfy", "fugashi": "fugashi>=1.0", "GitPython": "GitPython<3.1.19", "hf-doc-builder": "hf-doc-builder>=0.3.0", "huggingface-hub": "huggingface-hub>=0.14.1,<1.0", "importlib_metadata": "importlib_metadata", "ipadic": "ipadic>=1.0.0,<2.0", "isort": "isort>=5.5.4", "jax": "jax>=0.2.8,!=0.3.2,<=0.4.13", "jaxlib": "jaxlib>=0.1.65,<=0.4.13", "jieba": "jieba", "kenlm": "kenlm", "keras-nlp": "keras-nlp>=0.3.1", "librosa": "librosa", "nltk": "nltk", "natten": "natten>=0.14.6", "numpy": "numpy>=1.17", "onnxconverter-common": "onnxconverter-common", "onnxruntime-tools": "onnxruntime-tools>=1.4.2", "onnxruntime": "onnxruntime>=1.4.0", "opencv-python": "opencv-python", "optuna": "optuna", "optax": "optax>=0.0.8,<=0.1.4", "packaging": "packaging>=20.0", "parameterized": "parameterized", "phonemizer": "phonemizer", "protobuf": "protobuf", "psutil": "psutil", "pyyaml": "pyyaml>=5.1", "pydantic": "pydantic<2", "pytest": "pytest>=7.2.0", "pytest-timeout": "pytest-timeout", "pytest-xdist": "pytest-xdist", "python": "python>=3.8.0", "ray[tune]": "ray[tune]", "regex": "regex!=2019.12.17", "requests": "requests", "rhoknp": "rhoknp>=1.1.0,<1.3.1", "rjieba": "rjieba", "rouge-score": "rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1", "ruff": "ruff>=0.0.241,<=0.0.259", "sacrebleu": "sacrebleu>=1.4.12,<2.0.0", "sacremoses": "sacremoses", "safetensors": "safetensors>=0.3.1", "sagemaker": "sagemaker>=2.31.0", "scikit-learn": "scikit-learn", "sentencepiece": "sentencepiece>=0.1.91,!=0.1.92", "sigopt": "sigopt", "starlette": "starlette", "sudachipy": "sudachipy>=0.6.6", "sudachidict_core": "sudachidict_core>=20220729", "tensorflow-cpu": "tensorflow-cpu>=2.6,<2.14", "tensorflow": "tensorflow>=2.6,<2.14", "tensorflow-text": "tensorflow-text<2.14", "tf2onnx": "tf2onnx", "timeout-decorator": "timeout-decorator", "timm": "timm", "tokenizers": "tokenizers>=0.11.1,!=0.11.3,<0.14", "torch": "torch>=1.9,!=1.12.0", "torchaudio": "torchaudio", "torchvision": "torchvision", "pyctcdecode": "pyctcdecode>=0.4.0", "tqdm": "tqdm>=4.27", "unidic": "unidic>=1.0.2", "unidic_lite": "unidic_lite>=1.0.7", "urllib3": "urllib3<2.0.0", "uvicorn": "uvicorn", }
625
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowercase : Tuple = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : int = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys _lowercase : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
625
"""simple docstring""" def snake_case__ ( __lowerCamelCase : list[int] ): """simple docstring""" if not numbers: return 0 if not isinstance(__lowerCamelCase , (list, tuple) ) or not all( isinstance(__lowerCamelCase , __lowerCamelCase ) for number in numbers ): raise ValueError('''numbers must be an iterable of integers''' ) lowerCamelCase__ : Any =numbers[0] for i in range(1 , len(__lowerCamelCase ) ): # update the maximum and minimum subarray products lowerCamelCase__ : Dict =numbers[i] if number < 0: lowerCamelCase__ , lowerCamelCase__ : List[Any] =min_till_now, max_till_now lowerCamelCase__ : Optional[int] =max(__lowerCamelCase , max_till_now * number ) lowerCamelCase__ : Dict =min(__lowerCamelCase , min_till_now * number ) # update the maximum product found till now lowerCamelCase__ : Tuple =max(__lowerCamelCase , __lowerCamelCase ) return max_prod
625
1
"""simple docstring""" import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self : Any, lowerCamelCase : Optional[Any]=0.01, lowerCamelCase : Any=1000 )-> Any: lowerCamelCase__ : int =p_stop lowerCamelCase__ : int =max_length def __iter__( self : List[Any] )-> Any: lowerCamelCase__ : Dict =0 lowerCamelCase__ : List[Any] =False while not stop and count < self.max_length: yield count count += 1 lowerCamelCase__ : Dict =random.random() < self.p_stop class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def snake_case ( self : Optional[int], lowerCamelCase : Tuple, lowerCamelCase : Union[str, Any], lowerCamelCase : str=False, lowerCamelCase : Optional[Any]=True )-> Union[str, Any]: lowerCamelCase__ : Tuple =[ BatchSamplerShard(lowerCamelCase, 2, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) for i in range(2 ) ] lowerCamelCase__ : Any =[list(lowerCamelCase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowerCamelCase ) for shard in batch_sampler_shards], [len(lowerCamelCase ) for e in expected] ) self.assertListEqual(lowerCamelCase, lowerCamelCase ) def snake_case ( self : Tuple )-> Any: # Check the shards when the dataset is a round multiple of total batch size. lowerCamelCase__ : Optional[int] =BatchSampler(range(24 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : List[str] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Tuple =BatchSampler(range(24 ), batch_size=3, drop_last=lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowerCamelCase__ : List[Any] =BatchSampler(range(21 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : List[str] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : str =BatchSampler(range(21 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowerCamelCase__ : Dict =BatchSampler(range(22 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[int] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : List[str] =BatchSampler(range(22 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[int] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowerCamelCase__ : Optional[int] =BatchSampler(range(20 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : int =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : List[Any] =BatchSampler(range(20 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : str =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) # Check the shards when the dataset is very small. lowerCamelCase__ : Optional[int] =BatchSampler(range(2 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[int] =[[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Optional[Any] =BatchSampler(range(2 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[Any] =[[], []] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase ) def snake_case ( self : List[str] )-> Union[str, Any]: # Check the shards when the dataset is a round multiple of batch size. lowerCamelCase__ : List[Any] =BatchSampler(range(24 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : List[Any] =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) lowerCamelCase__ : str =BatchSampler(range(24 ), batch_size=4, drop_last=lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. lowerCamelCase__ : Union[str, Any] =BatchSampler(range(22 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Any =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) lowerCamelCase__ : List[str] =BatchSampler(range(22 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : str =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowerCamelCase__ : Dict =BatchSampler(range(21 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : List[Any] =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) lowerCamelCase__ : str =BatchSampler(range(21 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Tuple =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) # Check the shards when the dataset is very small. lowerCamelCase__ : Any =BatchSampler(range(2 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Tuple =[[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =BatchSampler(range(2 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[int] =[[], []] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase ) def snake_case ( self : Optional[Any] )-> Tuple: # Check the shards when the dataset is a round multiple of total batch size. lowerCamelCase__ : Dict =BatchSampler(range(24 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[int] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : Any =BatchSampler(range(24 ), batch_size=3, drop_last=lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowerCamelCase__ : Optional[Any] =BatchSampler(range(21 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : str =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : List[Any] =BatchSampler(range(21 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Tuple =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowerCamelCase__ : str =BatchSampler(range(22 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : str =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : List[str] =BatchSampler(range(22 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Dict =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowerCamelCase__ : int =BatchSampler(range(20 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : List[str] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : List[str] =BatchSampler(range(20 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[Any] =[ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is very small. lowerCamelCase__ : Dict =BatchSampler(range(2 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : int =[[[0, 1]], []] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : str =BatchSampler(range(2 ), batch_size=3, drop_last=lowerCamelCase ) lowerCamelCase__ : int =[[], []] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, even_batches=lowerCamelCase ) def snake_case ( self : Tuple )-> List[str]: # Check the shards when the dataset is a round multiple of batch size. lowerCamelCase__ : Tuple =BatchSampler(range(24 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[int] =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : List[str] =BatchSampler(range(24 ), batch_size=4, drop_last=lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. lowerCamelCase__ : Tuple =BatchSampler(range(22 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : List[Any] =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : Tuple =BatchSampler(range(22 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowerCamelCase__ : Optional[int] =BatchSampler(range(21 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : str =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : int =BatchSampler(range(21 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Optional[Any] =[ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) # Check the shards when the dataset is very small. lowerCamelCase__ : Dict =BatchSampler(range(2 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Any =[[[0, 1]], []] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) lowerCamelCase__ : Any =BatchSampler(range(2 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : Any =[[], []] self.check_batch_sampler_shards(lowerCamelCase, lowerCamelCase, split_batches=lowerCamelCase, even_batches=lowerCamelCase ) def snake_case ( self : List[str] )-> Dict: lowerCamelCase__ : Dict =[[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] lowerCamelCase__ : List[Any] =[BatchSamplerShard(lowerCamelCase, 2, lowerCamelCase, even_batches=lowerCamelCase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ), 3 ) self.assertEqual(len(batch_sampler_shards[1] ), 2 ) self.assertListEqual(list(batch_sampler_shards[0] ), [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ), [[3, 4], [9, 10, 11]] ) def snake_case ( self : Optional[int], lowerCamelCase : Any, lowerCamelCase : Dict, lowerCamelCase : Any, lowerCamelCase : List[Any]=False, lowerCamelCase : Optional[int]=2, lowerCamelCase : Any=False )-> Optional[int]: random.seed(lowerCamelCase ) lowerCamelCase__ : int =list(lowerCamelCase ) lowerCamelCase__ : List[str] =[ IterableDatasetShard( lowerCamelCase, batch_size=lowerCamelCase, drop_last=lowerCamelCase, num_processes=lowerCamelCase, process_index=lowerCamelCase, split_batches=lowerCamelCase, ) for i in range(lowerCamelCase ) ] lowerCamelCase__ : Optional[int] =[] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowerCamelCase ) iterable_dataset_lists.append(list(lowerCamelCase ) ) lowerCamelCase__ : Union[str, Any] =batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size lowerCamelCase__ : Any =iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowerCamelCase ), len(lowerCamelCase ) ) self.assertTrue(len(lowerCamelCase ) % shard_batch_size == 0 ) lowerCamelCase__ : Tuple =[] for idx in range(0, len(lowerCamelCase ), lowerCamelCase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowerCamelCase ) < len(lowerCamelCase ): reference += reference self.assertListEqual(lowerCamelCase, reference[: len(lowerCamelCase )] ) def snake_case ( self : List[Any] )-> Any: lowerCamelCase__ : Optional[int] =42 lowerCamelCase__ : List[str] =RandomIterableDataset() self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) # Edge case with a very small dataset lowerCamelCase__ : Optional[Any] =RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) self.check_iterable_dataset_shards(lowerCamelCase, lowerCamelCase, batch_size=4, drop_last=lowerCamelCase, split_batches=lowerCamelCase ) def snake_case ( self : int )-> List[Any]: lowerCamelCase__ : Optional[Any] =BatchSampler(range(16 ), batch_size=4, drop_last=lowerCamelCase ) lowerCamelCase__ : List[Any] =SkipBatchSampler(lowerCamelCase, 2 ) self.assertListEqual(list(lowerCamelCase ), [[8, 9, 10, 11], [12, 13, 14, 15]] ) def snake_case ( self : int )-> int: lowerCamelCase__ : List[str] =SkipDataLoader(list(range(16 ) ), batch_size=4, skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader], [[8, 9, 10, 11], [12, 13, 14, 15]] ) def snake_case ( self : Tuple )-> Dict: lowerCamelCase__ : Any =DataLoader(list(range(16 ) ), batch_size=4 ) lowerCamelCase__ : int =skip_first_batches(lowerCamelCase, num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader], [[8, 9, 10, 11], [12, 13, 14, 15]] ) def snake_case ( self : List[Any] )-> Any: lowerCamelCase__ : Dict =DataLoaderShard(list(range(16 ) ), batch_size=4 ) for idx, _ in enumerate(lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) def snake_case ( self : Optional[int] )-> Tuple: Accelerator() lowerCamelCase__ : List[Any] =DataLoaderDispatcher(range(16 ), batch_size=4 ) for idx, _ in enumerate(lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 )
625
"""simple docstring""" from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 42 _a = 42 class __SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' _a = 42 _a = (1_6, 3_2, 9_6, 2_5_6) _a = jnp.floataa def snake_case ( self : Tuple )-> int: lowerCamelCase__ : Tuple =nn.Conv( self.block_out_channels[0], kernel_size=(3, 3), padding=((1, 1), (1, 1)), dtype=self.dtype, ) lowerCamelCase__ : Dict =[] for i in range(len(self.block_out_channels ) - 1 ): lowerCamelCase__ : Dict =self.block_out_channels[i] lowerCamelCase__ : Dict =self.block_out_channels[i + 1] lowerCamelCase__ : List[str] =nn.Conv( lowerCamelCase, kernel_size=(3, 3), padding=((1, 1), (1, 1)), dtype=self.dtype, ) blocks.append(lowerCamelCase ) lowerCamelCase__ : Optional[int] =nn.Conv( lowerCamelCase, kernel_size=(3, 3), strides=(2, 2), padding=((1, 1), (1, 1)), dtype=self.dtype, ) blocks.append(lowerCamelCase ) lowerCamelCase__ : Any =blocks lowerCamelCase__ : Optional[int] =nn.Conv( self.conditioning_embedding_channels, kernel_size=(3, 3), padding=((1, 1), (1, 1)), kernel_init=nn.initializers.zeros_init(), bias_init=nn.initializers.zeros_init(), dtype=self.dtype, ) def __call__( self : Any, lowerCamelCase : int )-> List[str]: lowerCamelCase__ : Tuple =self.conv_in(lowerCamelCase ) lowerCamelCase__ : Dict =nn.silu(lowerCamelCase ) for block in self.blocks: lowerCamelCase__ : str =block(lowerCamelCase ) lowerCamelCase__ : List[str] =nn.silu(lowerCamelCase ) lowerCamelCase__ : Any =self.conv_out(lowerCamelCase ) return embedding @flax_register_to_config class __SCREAMING_SNAKE_CASE ( nn.Module , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' _a = 3_2 _a = 4 _a = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) _a = False _a = (3_2_0, 6_4_0, 1_2_8_0, 1_2_8_0) _a = 2 _a = 8 _a = None _a = 1_2_8_0 _a = 0.0 _a = False _a = jnp.floataa _a = True _a = 0 _a = "rgb" _a = (1_6, 3_2, 9_6, 2_5_6) def snake_case ( self : str, lowerCamelCase : jax.random.KeyArray )-> FrozenDict: # init input tensors lowerCamelCase__ : int =(1, self.in_channels, self.sample_size, self.sample_size) lowerCamelCase__ : int =jnp.zeros(lowerCamelCase, dtype=jnp.floataa ) lowerCamelCase__ : Union[str, Any] =jnp.ones((1,), dtype=jnp.intaa ) lowerCamelCase__ : str =jnp.zeros((1, 1, self.cross_attention_dim), dtype=jnp.floataa ) lowerCamelCase__ : Any =(1, 3, self.sample_size * 8, self.sample_size * 8) lowerCamelCase__ : Optional[Any] =jnp.zeros(lowerCamelCase, dtype=jnp.floataa ) lowerCamelCase__ , lowerCamelCase__ : List[Any] =jax.random.split(lowerCamelCase ) lowerCamelCase__ : Dict ={'''params''': params_rng, '''dropout''': dropout_rng} return self.init(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase )["params"] def snake_case ( self : Any )-> Tuple: lowerCamelCase__ : Optional[int] =self.block_out_channels lowerCamelCase__ : Tuple =block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. lowerCamelCase__ : List[Any] =self.num_attention_heads or self.attention_head_dim # input lowerCamelCase__ : int =nn.Conv( block_out_channels[0], kernel_size=(3, 3), strides=(1, 1), padding=((1, 1), (1, 1)), dtype=self.dtype, ) # time lowerCamelCase__ : str =FlaxTimesteps( block_out_channels[0], flip_sin_to_cos=self.flip_sin_to_cos, freq_shift=self.config.freq_shift ) lowerCamelCase__ : Dict =FlaxTimestepEmbedding(lowerCamelCase, dtype=self.dtype ) lowerCamelCase__ : List[Any] =FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0], block_out_channels=self.conditioning_embedding_out_channels, ) lowerCamelCase__ : Dict =self.only_cross_attention if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase__ : int =(only_cross_attention,) * len(self.down_block_types ) if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase__ : List[str] =(num_attention_heads,) * len(self.down_block_types ) # down lowerCamelCase__ : Union[str, Any] =[] lowerCamelCase__ : Dict =[] lowerCamelCase__ : List[Any] =block_out_channels[0] lowerCamelCase__ : List[Any] =nn.Conv( lowerCamelCase, kernel_size=(1, 1), padding='''VALID''', kernel_init=nn.initializers.zeros_init(), bias_init=nn.initializers.zeros_init(), dtype=self.dtype, ) controlnet_down_blocks.append(lowerCamelCase ) for i, down_block_type in enumerate(self.down_block_types ): lowerCamelCase__ : List[Any] =output_channel lowerCamelCase__ : str =block_out_channels[i] lowerCamelCase__ : Dict =i == len(lowerCamelCase ) - 1 if down_block_type == "CrossAttnDownBlock2D": lowerCamelCase__ : str =FlaxCrossAttnDownBlockaD( in_channels=lowerCamelCase, out_channels=lowerCamelCase, dropout=self.dropout, num_layers=self.layers_per_block, num_attention_heads=num_attention_heads[i], add_downsample=not is_final_block, use_linear_projection=self.use_linear_projection, only_cross_attention=only_cross_attention[i], dtype=self.dtype, ) else: lowerCamelCase__ : List[Any] =FlaxDownBlockaD( in_channels=lowerCamelCase, out_channels=lowerCamelCase, dropout=self.dropout, num_layers=self.layers_per_block, add_downsample=not is_final_block, dtype=self.dtype, ) down_blocks.append(lowerCamelCase ) for _ in range(self.layers_per_block ): lowerCamelCase__ : Any =nn.Conv( lowerCamelCase, kernel_size=(1, 1), padding='''VALID''', kernel_init=nn.initializers.zeros_init(), bias_init=nn.initializers.zeros_init(), dtype=self.dtype, ) controlnet_down_blocks.append(lowerCamelCase ) if not is_final_block: lowerCamelCase__ : Any =nn.Conv( lowerCamelCase, kernel_size=(1, 1), padding='''VALID''', kernel_init=nn.initializers.zeros_init(), bias_init=nn.initializers.zeros_init(), dtype=self.dtype, ) controlnet_down_blocks.append(lowerCamelCase ) lowerCamelCase__ : int =down_blocks lowerCamelCase__ : List[str] =controlnet_down_blocks # mid lowerCamelCase__ : Tuple =block_out_channels[-1] lowerCamelCase__ : List[Any] =FlaxUNetMidBlockaDCrossAttn( in_channels=lowerCamelCase, dropout=self.dropout, num_attention_heads=num_attention_heads[-1], use_linear_projection=self.use_linear_projection, dtype=self.dtype, ) lowerCamelCase__ : List[str] =nn.Conv( lowerCamelCase, kernel_size=(1, 1), padding='''VALID''', kernel_init=nn.initializers.zeros_init(), bias_init=nn.initializers.zeros_init(), dtype=self.dtype, ) def __call__( self : int, lowerCamelCase : List[Any], lowerCamelCase : Tuple, lowerCamelCase : Union[str, Any], lowerCamelCase : str, lowerCamelCase : float = 1.0, lowerCamelCase : bool = True, lowerCamelCase : bool = False, )-> Union[FlaxControlNetOutput, Tuple]: lowerCamelCase__ : int =self.controlnet_conditioning_channel_order if channel_order == "bgr": lowerCamelCase__ : int =jnp.flip(lowerCamelCase, axis=1 ) # 1. time if not isinstance(lowerCamelCase, jnp.ndarray ): lowerCamelCase__ : Any =jnp.array([timesteps], dtype=jnp.intaa ) elif isinstance(lowerCamelCase, jnp.ndarray ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] =timesteps.astype(dtype=jnp.floataa ) lowerCamelCase__ : int =jnp.expand_dims(lowerCamelCase, 0 ) lowerCamelCase__ : Optional[Any] =self.time_proj(lowerCamelCase ) lowerCamelCase__ : Optional[Any] =self.time_embedding(lowerCamelCase ) # 2. pre-process lowerCamelCase__ : Optional[int] =jnp.transpose(lowerCamelCase, (0, 2, 3, 1) ) lowerCamelCase__ : Dict =self.conv_in(lowerCamelCase ) lowerCamelCase__ : List[str] =jnp.transpose(lowerCamelCase, (0, 2, 3, 1) ) lowerCamelCase__ : int =self.controlnet_cond_embedding(lowerCamelCase ) sample += controlnet_cond # 3. down lowerCamelCase__ : Union[str, Any] =(sample,) for down_block in self.down_blocks: if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase__ , lowerCamelCase__ : Dict =down_block(lowerCamelCase, lowerCamelCase, lowerCamelCase, deterministic=not train ) else: lowerCamelCase__ , lowerCamelCase__ : Tuple =down_block(lowerCamelCase, lowerCamelCase, deterministic=not train ) down_block_res_samples += res_samples # 4. mid lowerCamelCase__ : Optional[int] =self.mid_block(lowerCamelCase, lowerCamelCase, lowerCamelCase, deterministic=not train ) # 5. contronet blocks lowerCamelCase__ : Optional[Any] =() for down_block_res_sample, controlnet_block in zip(lowerCamelCase, self.controlnet_down_blocks ): lowerCamelCase__ : Union[str, Any] =controlnet_block(lowerCamelCase ) controlnet_down_block_res_samples += (down_block_res_sample,) lowerCamelCase__ : List[str] =controlnet_down_block_res_samples lowerCamelCase__ : List[str] =self.controlnet_mid_block(lowerCamelCase ) # 6. scaling lowerCamelCase__ : Union[str, Any] =[sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=lowerCamelCase, mid_block_res_sample=lowerCamelCase )
625
1
"""simple docstring""" # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _lowercase : Dict = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model _lowercase : List[str] = { # fairseq: "wmt19-ru-en": {"length_penalty": 1.1}, "wmt19-en-ru": {"length_penalty": 1.15}, "wmt19-en-de": {"length_penalty": 1.0}, "wmt19-de-en": {"length_penalty": 1.1}, # allenai: "wmt16-en-de-dist-12-1": {"length_penalty": 0.6}, "wmt16-en-de-dist-6-1": {"length_penalty": 0.6}, "wmt16-en-de-12-1": {"length_penalty": 0.8}, "wmt19-de-en-6-6-base": {"length_penalty": 0.6}, "wmt19-de-en-6-6-big": {"length_penalty": 0.6}, } # this remaps the different models to their organization names _lowercase : int = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _lowercase : List[Any] = "facebook" for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: _lowercase : Optional[Any] = "allenai" def snake_case__ ( __lowerCamelCase : List[str] ): """simple docstring""" # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} lowerCamelCase__ : Any =dict((re.sub(R'''@@$''' , '''''' , __lowerCamelCase ), v) if k.endswith('''@@''' ) else (re.sub(R'''$''' , '''</w>''' , __lowerCamelCase ), v) for k, v in d.items() ) lowerCamelCase__ : Dict ='''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[f'''{k}</w>'''] lowerCamelCase__ : int =d[k] # restore return da def snake_case__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] ): """simple docstring""" # prep assert os.path.exists(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) print(f'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models lowerCamelCase__ : List[str] =basename(__lowerCamelCase ) lowerCamelCase__ : Tuple =dirname(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =fairseq.model_parallel.models.transformer.ModelParallelTransformerModel lowerCamelCase__ : Any =cls.hub_models() lowerCamelCase__ : int ={'''bpe''': '''fastbpe''', '''tokenizer''': '''moses'''} lowerCamelCase__ : Dict ='''.''' # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(f'''using checkpoint {checkpoint_file}''' ) lowerCamelCase__ : Tuple =hub_utils.from_pretrained( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , archive_map=__lowerCamelCase , **__lowerCamelCase ) lowerCamelCase__ : int =vars(chkpt['''args''']['''model'''] ) lowerCamelCase__ : Dict =args['''source_lang'''] lowerCamelCase__ : List[str] =args['''target_lang'''] lowerCamelCase__ : Union[str, Any] =dirname(__lowerCamelCase ) lowerCamelCase__ : Optional[int] =basename(__lowerCamelCase ) # dicts lowerCamelCase__ : str =os.path.join(__lowerCamelCase , f'''dict.{src_lang}.txt''' ) lowerCamelCase__ : Tuple =os.path.join(__lowerCamelCase , f'''dict.{tgt_lang}.txt''' ) lowerCamelCase__ : Any =Dictionary.load(__lowerCamelCase ) lowerCamelCase__ : Tuple =rewrite_dict_keys(src_dict.indices ) lowerCamelCase__ : Dict =len(__lowerCamelCase ) lowerCamelCase__ : Dict =os.path.join(__lowerCamelCase , '''vocab-src.json''' ) print(f'''Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records''' ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab lowerCamelCase__ : Tuple =True for k in src_vocab.keys(): if not k.islower(): lowerCamelCase__ : Union[str, Any] =False break lowerCamelCase__ : List[Any] =Dictionary.load(__lowerCamelCase ) lowerCamelCase__ : Any =rewrite_dict_keys(tgt_dict.indices ) lowerCamelCase__ : Union[str, Any] =len(__lowerCamelCase ) lowerCamelCase__ : Optional[Any] =os.path.join(__lowerCamelCase , '''vocab-tgt.json''' ) print(f'''Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records''' ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # merges_file (bpecodes) lowerCamelCase__ : Optional[Any] =os.path.join(__lowerCamelCase , VOCAB_FILES_NAMES['''merges_file'''] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" lowerCamelCase__ : List[Any] =os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): break with open(__lowerCamelCase , encoding='''utf-8''' ) as fin: lowerCamelCase__ : Union[str, Any] =fin.read() lowerCamelCase__ : Any =re.sub(R''' \d+$''' , '''''' , __lowerCamelCase , 0 , re.M ) # remove frequency number print(f'''Generating {merges_file}''' ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' ) as fout: fout.write(__lowerCamelCase ) # model config lowerCamelCase__ : Union[str, Any] =os.path.join(__lowerCamelCase , '''config.json''' ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", f'''need to extend tokenizer to support bpe={args["bpe"]}''' assert args["tokenizer"] == "moses", f'''need to extend tokenizer to support bpe={args["tokenizer"]}''' lowerCamelCase__ : Optional[int] ={ '''architectures''': ['''FSMTForConditionalGeneration'''], '''model_type''': '''fsmt''', '''activation_dropout''': args['''activation_dropout'''], '''activation_function''': '''relu''', '''attention_dropout''': args['''attention_dropout'''], '''d_model''': args['''decoder_embed_dim'''], '''dropout''': args['''dropout'''], '''init_std''': 0.02, '''max_position_embeddings''': args['''max_source_positions'''], '''num_hidden_layers''': args['''encoder_layers'''], '''src_vocab_size''': src_vocab_size, '''tgt_vocab_size''': tgt_vocab_size, '''langs''': [src_lang, tgt_lang], '''encoder_attention_heads''': args['''encoder_attention_heads'''], '''encoder_ffn_dim''': args['''encoder_ffn_embed_dim'''], '''encoder_layerdrop''': args['''encoder_layerdrop'''], '''encoder_layers''': args['''encoder_layers'''], '''decoder_attention_heads''': args['''decoder_attention_heads'''], '''decoder_ffn_dim''': args['''decoder_ffn_embed_dim'''], '''decoder_layerdrop''': args['''decoder_layerdrop'''], '''decoder_layers''': args['''decoder_layers'''], '''bos_token_id''': 0, '''pad_token_id''': 1, '''eos_token_id''': 2, '''is_encoder_decoder''': True, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_all_embeddings'''], } # good hparam defaults to start with lowerCamelCase__ : Dict =5 lowerCamelCase__ : Tuple =False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: lowerCamelCase__ : Any =best_score_hparams[model_dir]['''length_penalty'''] else: lowerCamelCase__ : Tuple =1.0 print(f'''Generating {fsmt_model_config_file}''' ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # tokenizer config lowerCamelCase__ : Optional[Any] =os.path.join(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : Optional[int] ={ '''langs''': [src_lang, tgt_lang], '''model_max_length''': 1024, '''do_lower_case''': do_lower_case, } print(f'''Generating {fsmt_tokenizer_config_file}''' ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # model lowerCamelCase__ : Tuple =chkpt['''models'''][0] lowerCamelCase__ : Tuple =model.state_dict() # rename keys to start with 'model.' lowerCamelCase__ : Dict =OrderedDict(('''model.''' + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys lowerCamelCase__ : int =[ '''model.model''', '''model.encoder.version''', '''model.decoder.version''', '''model.encoder_embed_tokens.weight''', '''model.decoder_embed_tokens.weight''', '''model.encoder.embed_positions._float_tensor''', '''model.decoder.embed_positions._float_tensor''', ] for k in ignore_keys: model_state_dict.pop(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : Tuple =FSMTConfig.from_pretrained(__lowerCamelCase ) lowerCamelCase__ : Tuple =FSMTForConditionalGeneration(__lowerCamelCase ) # check that it loads ok model_new.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) # save lowerCamelCase__ : Tuple =os.path.join(__lowerCamelCase , __lowerCamelCase ) print(f'''Generating {pytorch_weights_dump_path}''' ) torch.save(__lowerCamelCase , __lowerCamelCase ) print('''Conversion is done!''' ) print('''\nLast step is to upload the files to s3''' ) print(f'''cd {data_root}''' ) print(f'''transformers-cli upload {model_dir}''' ) if __name__ == "__main__": _lowercase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--fsmt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _lowercase : str = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
625
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowercase : Optional[Any] = { "configuration_clip": [ "CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "CLIPConfig", "CLIPOnnxConfig", "CLIPTextConfig", "CLIPVisionConfig", ], "processing_clip": ["CLIPProcessor"], "tokenization_clip": ["CLIPTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : str = ["CLIPTokenizerFast"] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Any = ["CLIPFeatureExtractor"] _lowercase : int = ["CLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[Any] = [ "CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "CLIPModel", "CLIPPreTrainedModel", "CLIPTextModel", "CLIPTextModelWithProjection", "CLIPVisionModel", "CLIPVisionModelWithProjection", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Dict = [ "TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCLIPModel", "TFCLIPPreTrainedModel", "TFCLIPTextModel", "TFCLIPVisionModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Union[str, Any] = [ "FlaxCLIPModel", "FlaxCLIPPreTrainedModel", "FlaxCLIPTextModel", "FlaxCLIPTextPreTrainedModel", "FlaxCLIPVisionModel", "FlaxCLIPVisionPreTrainedModel", ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys _lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
625
1
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) _lowercase : List[Any] = "hf-internal-testing/tiny-random-bert" _lowercase : List[Any] = os.path.join(TRANSFORMERS_CACHE, "models--hf-internal-testing--tiny-random-bert") _lowercase : str = "9b8c223d42b2188cb49d29af482996f9d0f3e5a6" class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def snake_case ( self : int )-> Union[str, Any]: lowerCamelCase__ : Dict =cached_file(lowerCamelCase, lowerCamelCase ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(lowerCamelCase ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(lowerCamelCase, lowerCamelCase ) ) ) with open(os.path.join(lowerCamelCase, '''refs''', '''main''' ) ) as f: lowerCamelCase__ : Union[str, Any] =f.read() self.assertEqual(lowerCamelCase, os.path.join(lowerCamelCase, '''snapshots''', lowerCamelCase, lowerCamelCase ) ) self.assertTrue(os.path.isfile(lowerCamelCase ) ) # File is cached at the same place the second time. lowerCamelCase__ : str =cached_file(lowerCamelCase, lowerCamelCase ) self.assertEqual(lowerCamelCase, lowerCamelCase ) # Using a specific revision to test the full commit hash. lowerCamelCase__ : Dict =cached_file(lowerCamelCase, lowerCamelCase, revision='''9b8c223''' ) self.assertEqual(lowerCamelCase, os.path.join(lowerCamelCase, '''snapshots''', lowerCamelCase, lowerCamelCase ) ) def snake_case ( self : Optional[Any] )-> List[str]: with self.assertRaisesRegex(lowerCamelCase, '''is not a valid model identifier''' ): lowerCamelCase__ : List[str] =cached_file('''tiny-random-bert''', lowerCamelCase ) with self.assertRaisesRegex(lowerCamelCase, '''is not a valid git identifier''' ): lowerCamelCase__ : Union[str, Any] =cached_file(lowerCamelCase, lowerCamelCase, revision='''aaaa''' ) with self.assertRaisesRegex(lowerCamelCase, '''does not appear to have a file named''' ): lowerCamelCase__ : int =cached_file(lowerCamelCase, '''conf''' ) def snake_case ( self : int )-> Optional[Any]: with self.assertRaisesRegex(lowerCamelCase, '''does not appear to have a file named''' ): lowerCamelCase__ : Dict =cached_file(lowerCamelCase, '''conf''' ) with open(os.path.join(lowerCamelCase, '''refs''', '''main''' ) ) as f: lowerCamelCase__ : Union[str, Any] =f.read() self.assertTrue(os.path.isfile(os.path.join(lowerCamelCase, '''.no_exist''', lowerCamelCase, '''conf''' ) ) ) lowerCamelCase__ : Optional[Any] =cached_file(lowerCamelCase, '''conf''', _raise_exceptions_for_missing_entries=lowerCamelCase ) self.assertIsNone(lowerCamelCase ) lowerCamelCase__ : List[Any] =cached_file(lowerCamelCase, '''conf''', local_files_only=lowerCamelCase, _raise_exceptions_for_missing_entries=lowerCamelCase ) self.assertIsNone(lowerCamelCase ) lowerCamelCase__ : Optional[Any] =mock.Mock() lowerCamelCase__ : int =500 lowerCamelCase__ : Any ={} lowerCamelCase__ : List[Any] =HTTPError lowerCamelCase__ : Optional[int] ={} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''', return_value=lowerCamelCase ) as mock_head: lowerCamelCase__ : Union[str, Any] =cached_file(lowerCamelCase, '''conf''', _raise_exceptions_for_connection_errors=lowerCamelCase ) self.assertIsNone(lowerCamelCase ) # This check we did call the fake head request mock_head.assert_called() def snake_case ( self : Optional[int] )-> int: self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''', lowerCamelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''', lowerCamelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''', lowerCamelCase ) ) def snake_case ( self : Optional[int] )-> Tuple: # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''', '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(lowerCamelCase, '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''', lowerCamelCase ) # The function raises if the revision does not exist. with self.assertRaisesRegex(lowerCamelCase, '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''', lowerCamelCase, revision='''ahaha''' ) lowerCamelCase__ : List[Any] =get_file_from_repo('''bert-base-cased''', lowerCamelCase ) # The name is the cached name which is not very easy to test, so instead we load the content. lowerCamelCase__ : Union[str, Any] =json.loads(open(lowerCamelCase, '''r''' ).read() ) self.assertEqual(config['''hidden_size'''], 768 ) def snake_case ( self : Tuple )-> Optional[int]: with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase__ : List[str] =Path(lowerCamelCase ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(lowerCamelCase, '''a.txt''' ), str(lowerCamelCase ) ) self.assertIsNone(get_file_from_repo(lowerCamelCase, '''b.txt''' ) )
625
"""simple docstring""" import os def snake_case__ ( ): """simple docstring""" with open(os.path.dirname(__lowerCamelCase ) + '''/p022_names.txt''' ) as file: lowerCamelCase__ : Tuple =str(file.readlines()[0] ) lowerCamelCase__ : int =names.replace('''"''' , '''''' ).split(''',''' ) names.sort() lowerCamelCase__ : Union[str, Any] =0 lowerCamelCase__ : str =0 for i, name in enumerate(__lowerCamelCase ): for letter in name: name_score += ord(__lowerCamelCase ) - 64 total_score += (i + 1) * name_score lowerCamelCase__ : Dict =0 return total_score if __name__ == "__main__": print(solution())
625
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowercase : List[str] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : str = ["OwlViTFeatureExtractor"] _lowercase : Dict = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Union[str, Any] = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _lowercase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
625
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : str, lowerCamelCase : int )-> None: lowerCamelCase__ : str =value lowerCamelCase__ : Node | None =None lowerCamelCase__ : Node | None =None class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : int, lowerCamelCase : Node )-> None: lowerCamelCase__ : Any =tree def snake_case ( self : str, lowerCamelCase : Node | None )-> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : Dict )-> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
625
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer _lowercase : Optional[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name _lowercase : Union[str, Any] = "\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n >>> repo = \"openai/shap-e-img2img\"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = \"https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png\"\n >>> image = load_image(image_url).convert(\"RGB\")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], \"corgi_3d.gif\")\n ```\n" @dataclass class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 42 class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self : Dict, lowerCamelCase : PriorTransformer, lowerCamelCase : CLIPVisionModel, lowerCamelCase : CLIPImageProcessor, lowerCamelCase : HeunDiscreteScheduler, lowerCamelCase : ShapERenderer, )-> str: super().__init__() self.register_modules( prior=lowerCamelCase, image_encoder=lowerCamelCase, image_processor=lowerCamelCase, scheduler=lowerCamelCase, renderer=lowerCamelCase, ) def snake_case ( self : List[str], lowerCamelCase : str, lowerCamelCase : List[Any], lowerCamelCase : List[str], lowerCamelCase : List[str], lowerCamelCase : Any, lowerCamelCase : List[str] )-> Optional[int]: if latents is None: lowerCamelCase__ : Dict =randn_tensor(lowerCamelCase, generator=lowerCamelCase, device=lowerCamelCase, dtype=lowerCamelCase ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowerCamelCase__ : List[str] =latents.to(lowerCamelCase ) lowerCamelCase__ : Dict =latents * scheduler.init_noise_sigma return latents def snake_case ( self : Optional[Any], lowerCamelCase : Union[str, Any]=0 )-> Optional[Any]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) lowerCamelCase__ : Optional[Any] =torch.device(F'''cuda:{gpu_id}''' ) lowerCamelCase__ : Tuple =[self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase, lowerCamelCase ) @property def snake_case ( self : Optional[Any] )-> Union[str, Any]: if self.device != torch.device('''meta''' ) or not hasattr(self.image_encoder, '''_hf_hook''' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(lowerCamelCase, '''_hf_hook''' ) and hasattr(module._hf_hook, '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def snake_case ( self : Optional[Any], lowerCamelCase : Any, lowerCamelCase : List[str], lowerCamelCase : Union[str, Any], lowerCamelCase : List[str], )-> Any: if isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(image[0], torch.Tensor ): lowerCamelCase__ : Dict =torch.cat(lowerCamelCase, axis=0 ) if image[0].ndim == 4 else torch.stack(lowerCamelCase, axis=0 ) if not isinstance(lowerCamelCase, torch.Tensor ): lowerCamelCase__ : str =self.image_processor(lowerCamelCase, return_tensors='''pt''' ).pixel_values[0].unsqueeze(0 ) lowerCamelCase__ : Any =image.to(dtype=self.image_encoder.dtype, device=lowerCamelCase ) lowerCamelCase__ : Any =self.image_encoder(lowerCamelCase )['''last_hidden_state'''] lowerCamelCase__ : Optional[int] =image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 lowerCamelCase__ : Dict =image_embeds.repeat_interleave(lowerCamelCase, dim=0 ) if do_classifier_free_guidance: lowerCamelCase__ : int =torch.zeros_like(lowerCamelCase ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowerCamelCase__ : List[str] =torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(lowerCamelCase ) def __call__( self : Tuple, lowerCamelCase : Union[PIL.Image.Image, List[PIL.Image.Image]], lowerCamelCase : int = 1, lowerCamelCase : int = 25, lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None, lowerCamelCase : Optional[torch.FloatTensor] = None, lowerCamelCase : float = 4.0, lowerCamelCase : int = 64, lowerCamelCase : Optional[str] = "pil", lowerCamelCase : bool = True, )-> Dict: if isinstance(lowerCamelCase, PIL.Image.Image ): lowerCamelCase__ : Dict =1 elif isinstance(lowerCamelCase, torch.Tensor ): lowerCamelCase__ : Union[str, Any] =image.shape[0] elif isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(image[0], (torch.Tensor, PIL.Image.Image) ): lowerCamelCase__ : List[Any] =len(lowerCamelCase ) else: raise ValueError( F'''`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(lowerCamelCase )}''' ) lowerCamelCase__ : Optional[Any] =self._execution_device lowerCamelCase__ : Optional[int] =batch_size * num_images_per_prompt lowerCamelCase__ : Any =guidance_scale > 1.0 lowerCamelCase__ : Dict =self._encode_image(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # prior self.scheduler.set_timesteps(lowerCamelCase, device=lowerCamelCase ) lowerCamelCase__ : Tuple =self.scheduler.timesteps lowerCamelCase__ : Optional[int] =self.prior.config.num_embeddings lowerCamelCase__ : Optional[Any] =self.prior.config.embedding_dim lowerCamelCase__ : Optional[Any] =self.prepare_latents( (batch_size, num_embeddings * embedding_dim), image_embeds.dtype, lowerCamelCase, lowerCamelCase, lowerCamelCase, self.scheduler, ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim lowerCamelCase__ : Any =latents.reshape(latents.shape[0], lowerCamelCase, lowerCamelCase ) for i, t in enumerate(self.progress_bar(lowerCamelCase ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase__ : int =torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase__ : Union[str, Any] =self.scheduler.scale_model_input(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ : Any =self.prior( lowerCamelCase, timestep=lowerCamelCase, proj_embedding=lowerCamelCase, ).predicted_image_embedding # remove the variance lowerCamelCase__ , lowerCamelCase__ : Optional[Any] =noise_pred.split( scaled_model_input.shape[2], dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: lowerCamelCase__ , lowerCamelCase__ : Dict =noise_pred.chunk(2 ) lowerCamelCase__ : Optional[int] =noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) lowerCamelCase__ : Optional[int] =self.scheduler.step( lowerCamelCase, timestep=lowerCamelCase, sample=lowerCamelCase, ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=lowerCamelCase ) lowerCamelCase__ : int =[] for i, latent in enumerate(lowerCamelCase ): print() lowerCamelCase__ : str =self.renderer.decode( latent[None, :], lowerCamelCase, size=lowerCamelCase, ray_batch_size=4096, n_coarse_samples=64, n_fine_samples=128, ) images.append(lowerCamelCase ) lowerCamelCase__ : List[Any] =torch.stack(lowerCamelCase ) if output_type not in ["np", "pil"]: raise ValueError(F'''Only the output types `pil` and `np` are supported not output_type={output_type}''' ) lowerCamelCase__ : Optional[int] =images.cpu().numpy() if output_type == "pil": lowerCamelCase__ : List[str] =[self.numpy_to_pil(lowerCamelCase ) for image in images] # Offload last model to CPU if hasattr(self, '''final_offload_hook''' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=lowerCamelCase )
625
"""simple docstring""" import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel _lowercase : List[str] = logging.getLogger(__name__) def snake_case__ ( __lowerCamelCase : Any , __lowerCamelCase : str ): """simple docstring""" # save results if os.path.exists(__lowerCamelCase ): if os.path.exists(os.path.join(__lowerCamelCase , '''config.json''' ) ) and os.path.isfile( os.path.join(__lowerCamelCase , '''config.json''' ) ): os.remove(os.path.join(__lowerCamelCase , '''config.json''' ) ) if os.path.exists(os.path.join(__lowerCamelCase , '''pytorch_model.bin''' ) ) and os.path.isfile( os.path.join(__lowerCamelCase , '''pytorch_model.bin''' ) ): os.remove(os.path.join(__lowerCamelCase , '''pytorch_model.bin''' ) ) else: os.makedirs(__lowerCamelCase ) model.save_pretrained(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Dict=False ): """simple docstring""" lowerCamelCase__ : Union[str, Any] =2 if unlogit: lowerCamelCase__ : Any =torch.pow(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : List[str] =p * torch.log(__lowerCamelCase ) lowerCamelCase__ : Tuple =0 return -plogp.sum(dim=-1 ) def snake_case__ ( __lowerCamelCase : Any ): """simple docstring""" logger.info('''lv, h >\t''' + '''\t'''.join(f'''{x + 1}''' for x in range(len(__lowerCamelCase ) ) ) ) for row in range(len(__lowerCamelCase ) ): if tensor.dtype != torch.long: logger.info(f'''layer {row + 1}:\t''' + '''\t'''.join(f'''{x:.5f}''' for x in tensor[row].cpu().data ) ) else: logger.info(f'''layer {row + 1}:\t''' + '''\t'''.join(f'''{x:d}''' for x in tensor[row].cpu().data ) ) def snake_case__ ( __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=True , __lowerCamelCase : List[Any]=True , __lowerCamelCase : List[str]=None , __lowerCamelCase : Tuple=False ): """simple docstring""" lowerCamelCase__ , lowerCamelCase__ : Tuple =model.config.num_hidden_layers, model.config.num_attention_heads lowerCamelCase__ : Optional[Any] =torch.zeros(__lowerCamelCase , __lowerCamelCase ).to(args.device ) lowerCamelCase__ : Optional[Any] =torch.zeros(__lowerCamelCase , __lowerCamelCase ).to(args.device ) if head_mask is None: lowerCamelCase__ : List[Any] =torch.ones(__lowerCamelCase , __lowerCamelCase ).to(args.device ) head_mask.requires_grad_(requires_grad=__lowerCamelCase ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: lowerCamelCase__ : Union[str, Any] =None lowerCamelCase__ : List[str] =0.0 lowerCamelCase__ : Union[str, Any] =0.0 for step, inputs in enumerate(tqdm(__lowerCamelCase , desc='''Iteration''' , disable=args.local_rank not in [-1, 0] ) ): lowerCamelCase__ : Any =tuple(t.to(args.device ) for t in inputs ) ((lowerCamelCase__) , ) : Any =inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) lowerCamelCase__ : Dict =model(__lowerCamelCase , labels=__lowerCamelCase , head_mask=__lowerCamelCase ) # (loss), lm_logits, presents, (all hidden_states), (attentions) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Tuple =( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(__lowerCamelCase ): lowerCamelCase__ : Any =entropy(attn.detach() , __lowerCamelCase ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(__lowerCamelCase ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: lowerCamelCase__ : int =2 lowerCamelCase__ : List[str] =torch.pow(torch.pow(__lowerCamelCase , __lowerCamelCase ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-2_0 if not args.dont_normalize_global_importance: lowerCamelCase__ : int =(head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info('''Attention entropies''' ) print_ad_tensor(__lowerCamelCase ) if compute_importance: logger.info('''Head importance scores''' ) print_ad_tensor(__lowerCamelCase ) logger.info('''Head ranked by importance scores''' ) lowerCamelCase__ : Optional[int] =torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) lowerCamelCase__ : Dict =torch.arange( head_importance.numel() , device=args.device ) lowerCamelCase__ : Any =head_ranks.view_as(__lowerCamelCase ) print_ad_tensor(__lowerCamelCase ) return attn_entropy, head_importance, total_loss def snake_case__ ( __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): """simple docstring""" lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Dict =compute_heads_importance(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , compute_entropy=__lowerCamelCase ) lowerCamelCase__ : int =1 / loss # instead of downsteam score use the LM loss logger.info('''Pruning: original score: %f, threshold: %f''' , __lowerCamelCase , original_score * args.masking_threshold ) lowerCamelCase__ : Dict =torch.ones_like(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =max(1 , int(new_head_mask.numel() * args.masking_amount ) ) lowerCamelCase__ : List[Any] =original_score while current_score >= original_score * args.masking_threshold: lowerCamelCase__ : List[Any] =new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads lowerCamelCase__ : int =float('''Inf''' ) lowerCamelCase__ : Union[str, Any] =head_importance.view(-1 ).sort()[1] if len(__lowerCamelCase ) <= num_to_mask: print('''BREAK BY num_to_mask''' ) break # mask heads lowerCamelCase__ : List[str] =current_heads_to_mask[:num_to_mask] logger.info('''Heads to mask: %s''' , str(current_heads_to_mask.tolist() ) ) lowerCamelCase__ : Optional[int] =new_head_mask.view(-1 ) lowerCamelCase__ : Optional[Any] =0.0 lowerCamelCase__ : Dict =new_head_mask.view_as(__lowerCamelCase ) lowerCamelCase__ : Tuple =new_head_mask.clone().detach() print_ad_tensor(__lowerCamelCase ) # Compute metric and head importance again lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[Any] =compute_heads_importance( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , compute_entropy=__lowerCamelCase , head_mask=__lowerCamelCase ) lowerCamelCase__ : Any =1 / loss logger.info( '''Masking: current score: %f, remaining heads %d (%.1f percents)''' , __lowerCamelCase , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info('''Final head mask''' ) print_ad_tensor(__lowerCamelCase ) np.save(os.path.join(args.output_dir , '''head_mask.npy''' ) , head_mask.detach().cpu().numpy() ) return head_mask def snake_case__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] ): """simple docstring""" lowerCamelCase__ : str =datetime.now() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[str] =compute_heads_importance( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , compute_entropy=__lowerCamelCase , compute_importance=__lowerCamelCase , head_mask=__lowerCamelCase ) lowerCamelCase__ : Tuple =1 / loss lowerCamelCase__ : Optional[Any] =datetime.now() - before_time lowerCamelCase__ : int =sum(p.numel() for p in model.parameters() ) lowerCamelCase__ : Any ={ layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(__lowerCamelCase ) ) } for k, v in heads_to_prune.items(): if isinstance(__lowerCamelCase , __lowerCamelCase ): lowerCamelCase__ : Optional[int] =[ v, ] assert sum(len(__lowerCamelCase ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(__lowerCamelCase ) lowerCamelCase__ : List[str] =sum(p.numel() for p in model.parameters() ) lowerCamelCase__ : Any =datetime.now() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Dict =compute_heads_importance( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , compute_entropy=__lowerCamelCase , compute_importance=__lowerCamelCase , head_mask=__lowerCamelCase , actually_pruned=__lowerCamelCase , ) lowerCamelCase__ : str =1 / loss lowerCamelCase__ : Union[str, Any] =datetime.now() - before_time logger.info( '''Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)''' , __lowerCamelCase , __lowerCamelCase , pruned_num_params / original_num_params * 100 , ) logger.info('''Pruning: score with masking: %f score with pruning: %f''' , __lowerCamelCase , __lowerCamelCase ) logger.info('''Pruning: speed ratio (original timing / new timing): %f percents''' , original_time / new_time * 100 ) save_model(__lowerCamelCase , args.output_dir ) def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : Optional[int] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--data_dir''' , default=__lowerCamelCase , type=__lowerCamelCase , required=__lowerCamelCase , help='''The input data dir. Should contain the .tsv files (or other data files) for the task.''' , ) parser.add_argument( '''--model_name_or_path''' , default=__lowerCamelCase , type=__lowerCamelCase , required=__lowerCamelCase , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--output_dir''' , default=__lowerCamelCase , type=__lowerCamelCase , required=__lowerCamelCase , help='''The output directory where the model predictions and checkpoints will be written.''' , ) # Other parameters parser.add_argument( '''--config_name''' , default='''''' , type=__lowerCamelCase , help='''Pretrained config name or path if not the same as model_name_or_path''' , ) parser.add_argument( '''--tokenizer_name''' , default='''''' , type=__lowerCamelCase , help='''Pretrained tokenizer name or path if not the same as model_name_or_path''' , ) parser.add_argument( '''--cache_dir''' , default=__lowerCamelCase , type=__lowerCamelCase , help='''Where do you want to store the pre-trained models downloaded from s3''' , ) parser.add_argument( '''--data_subset''' , type=__lowerCamelCase , default=-1 , help='''If > 0: limit the data to a subset of data_subset instances.''' ) parser.add_argument( '''--overwrite_output_dir''' , action='''store_true''' , help='''Whether to overwrite data in output directory''' ) parser.add_argument( '''--overwrite_cache''' , action='''store_true''' , help='''Overwrite the cached training and evaluation sets''' ) parser.add_argument( '''--dont_normalize_importance_by_layer''' , action='''store_true''' , help='''Don\'t normalize importance score by layers''' ) parser.add_argument( '''--dont_normalize_global_importance''' , action='''store_true''' , help='''Don\'t normalize all importance scores between 0 and 1''' , ) parser.add_argument( '''--try_masking''' , action='''store_true''' , help='''Whether to try to mask head until a threshold of accuracy.''' ) parser.add_argument( '''--masking_threshold''' , default=0.9 , type=__lowerCamelCase , help='''masking threshold in term of metrics (stop masking when metric < threshold * original metric value).''' , ) parser.add_argument( '''--masking_amount''' , default=0.1 , type=__lowerCamelCase , help='''Amount to heads to masking at each masking step.''' ) parser.add_argument('''--metric_name''' , default='''acc''' , type=__lowerCamelCase , help='''Metric to use for head masking.''' ) parser.add_argument( '''--max_seq_length''' , default=128 , type=__lowerCamelCase , help=( '''The maximum total input sequence length after WordPiece tokenization. \n''' '''Sequences longer than this will be truncated, sequences shorter padded.''' ) , ) parser.add_argument('''--batch_size''' , default=1 , type=__lowerCamelCase , help='''Batch size.''' ) parser.add_argument('''--seed''' , type=__lowerCamelCase , default=42 ) parser.add_argument('''--local_rank''' , type=__lowerCamelCase , default=-1 , help='''local_rank for distributed training on gpus''' ) parser.add_argument('''--no_cuda''' , action='''store_true''' , help='''Whether not to use CUDA when available''' ) parser.add_argument('''--server_ip''' , type=__lowerCamelCase , default='''''' , help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' , type=__lowerCamelCase , default='''''' , help='''Can be used for distant debugging.''' ) lowerCamelCase__ : List[Any] =parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=__lowerCamelCase ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: lowerCamelCase__ : Dict =torch.device('''cuda''' if torch.cuda.is_available() and not args.no_cuda else '''cpu''' ) lowerCamelCase__ : Dict =0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) lowerCamelCase__ : str =torch.device('''cuda''' , args.local_rank ) lowerCamelCase__ : Any =1 torch.distributed.init_process_group(backend='''nccl''' ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info('''device: {} n_gpu: {}, distributed: {}'''.format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) lowerCamelCase__ : Union[str, Any] =GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: lowerCamelCase__ : List[Any] =nn.parallel.DistributedDataParallel( __lowerCamelCase , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=__lowerCamelCase ) elif args.n_gpu > 1: lowerCamelCase__ : int =nn.DataParallel(__lowerCamelCase ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=__lowerCamelCase ) torch.save(__lowerCamelCase , os.path.join(args.output_dir , '''run_args.bin''' ) ) logger.info('''Training/evaluation parameters %s''' , __lowerCamelCase ) # Prepare dataset lowerCamelCase__ : Union[str, Any] =np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) lowerCamelCase__ : Any =(torch.from_numpy(__lowerCamelCase ),) lowerCamelCase__ : List[Any] =TensorDataset(*__lowerCamelCase ) lowerCamelCase__ : List[str] =RandomSampler(__lowerCamelCase ) lowerCamelCase__ : Dict =DataLoader(__lowerCamelCase , sampler=__lowerCamelCase , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: lowerCamelCase__ : Optional[int] =mask_heads(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) prune_heads(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": main()
625
1
"""simple docstring""" import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : List[Any] ): """simple docstring""" # Initialise PyTorch model lowerCamelCase__ : Any =TaConfig.from_json_file(__lowerCamelCase ) print(f'''Building PyTorch model from configuration: {config}''' ) lowerCamelCase__ : Optional[int] =TaForConditionalGeneration(__lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_ta(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _lowercase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _lowercase : int = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
625
"""simple docstring""" import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def snake_case__ ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : Tuple ): """simple docstring""" lowerCamelCase__ : Union[str, Any] =AutoConfig.from_pretrained(__lowerCamelCase ) lowerCamelCase__ : Any =FlaxAutoModelForSeqaSeqLM.from_config(config=__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =checkpoints.load_tax_checkpoint(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] ='''wi_0''' in tax_model['''target''']['''encoder''']['''layers_0''']['''mlp'''] if config.model_type == "t5": lowerCamelCase__ : List[str] ='''SelfAttention''' if config.model_type == "longt5" and config.encoder_attention_type == "local": lowerCamelCase__ : List[Any] ='''LocalSelfAttention''' elif config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowerCamelCase__ : Optional[Any] ='''TransientGlobalSelfAttention''' else: raise ValueError( '''Given config is expected to have `model_type=\'t5\'`, or `model_type=\'longt5` with `encoder_attention_type`''' ''' attribute with a value from [\'local\', \'transient-global].''' ) # Encoder for layer_index in range(config.num_layers ): lowerCamelCase__ : List[Any] =f'''layers_{str(__lowerCamelCase )}''' # Self-Attention lowerCamelCase__ : List[str] =tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''key''']['''kernel'''] lowerCamelCase__ : Optional[int] =tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''out''']['''kernel'''] lowerCamelCase__ : List[str] =tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''query''']['''kernel'''] lowerCamelCase__ : List[Any] =tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''value''']['''kernel'''] # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowerCamelCase__ : str =tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''T5LayerNorm_0''']['''scale'''] # Layer Normalization lowerCamelCase__ : List[Any] =tax_model['''target''']['''encoder'''][layer_name]['''pre_attention_layer_norm''']['''scale'''] if split_mlp_wi: lowerCamelCase__ : Optional[Any] =tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi_0''']['''kernel'''] lowerCamelCase__ : Dict =tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi_1''']['''kernel'''] else: lowerCamelCase__ : List[str] =tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi''']['''kernel'''] lowerCamelCase__ : Union[str, Any] =tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wo''']['''kernel'''] # Layer Normalization lowerCamelCase__ : Tuple =tax_model['''target''']['''encoder'''][layer_name]['''pre_mlp_layer_norm''']['''scale'''] # Assigning lowerCamelCase__ : str =flax_model.params['''encoder''']['''block'''][str(__lowerCamelCase )]['''layer'''] lowerCamelCase__ : int =tax_attention_key lowerCamelCase__ : Optional[int] =tax_attention_out lowerCamelCase__ : List[Any] =tax_attention_query lowerCamelCase__ : Optional[Any] =tax_attention_value lowerCamelCase__ : List[str] =tax_attention_layer_norm # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowerCamelCase__ : Optional[int] =tax_global_layer_norm if split_mlp_wi: lowerCamelCase__ : Optional[int] =tax_mlp_wi_a lowerCamelCase__ : Optional[int] =tax_mlp_wi_a else: lowerCamelCase__ : Union[str, Any] =tax_mlp_wi lowerCamelCase__ : str =tax_mlp_wo lowerCamelCase__ : Optional[Any] =tax_mlp_layer_norm lowerCamelCase__ : Optional[int] =flax_model_encoder_layer_block # Only for layer 0: lowerCamelCase__ : Union[str, Any] =tax_model['''target''']['''encoder''']['''relpos_bias''']['''rel_embedding'''].T lowerCamelCase__ : str =tax_encoder_rel_embedding # Side/global relative position_bias + layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": lowerCamelCase__ : Optional[int] =tax_model['''target''']['''encoder''']['''side_relpos_bias''']['''rel_embedding'''].T lowerCamelCase__ : Optional[int] =tax_encoder_global_rel_embedding # Assigning lowerCamelCase__ : int =tax_model['''target''']['''encoder''']['''encoder_norm''']['''scale'''] lowerCamelCase__ : List[Any] =tax_encoder_norm # Decoder for layer_index in range(config.num_layers ): lowerCamelCase__ : Dict =f'''layers_{str(__lowerCamelCase )}''' # Self-Attention lowerCamelCase__ : Dict =tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''key''']['''kernel'''] lowerCamelCase__ : List[Any] =tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''out''']['''kernel'''] lowerCamelCase__ : Optional[int] =tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''query''']['''kernel'''] lowerCamelCase__ : Union[str, Any] =tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''value''']['''kernel'''] # Layer Normalization lowerCamelCase__ : int =tax_model['''target''']['''decoder'''][layer_name]['''pre_self_attention_layer_norm'''][ '''scale''' ] # Encoder-Decoder-Attention lowerCamelCase__ : int =tax_model['''target''']['''decoder'''][layer_name]['''encoder_decoder_attention'''] lowerCamelCase__ : List[Any] =tax_enc_dec_attention_module['''key''']['''kernel'''] lowerCamelCase__ : Any =tax_enc_dec_attention_module['''out''']['''kernel'''] lowerCamelCase__ : Dict =tax_enc_dec_attention_module['''query''']['''kernel'''] lowerCamelCase__ : List[str] =tax_enc_dec_attention_module['''value''']['''kernel'''] # Layer Normalization lowerCamelCase__ : Dict =tax_model['''target''']['''decoder'''][layer_name]['''pre_cross_attention_layer_norm''']['''scale'''] # MLP if split_mlp_wi: lowerCamelCase__ : str =tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi_0''']['''kernel'''] lowerCamelCase__ : Any =tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi_1''']['''kernel'''] else: lowerCamelCase__ : List[Any] =tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi''']['''kernel'''] lowerCamelCase__ : Optional[Any] =tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wo''']['''kernel'''] # Layer Normalization lowerCamelCase__ : str =tax_model['''target''']['''decoder'''][layer_name]['''pre_mlp_layer_norm''']['''scale'''] # Assigning lowerCamelCase__ : str =flax_model.params['''decoder''']['''block'''][str(__lowerCamelCase )]['''layer'''] lowerCamelCase__ : Union[str, Any] =tax_attention_key lowerCamelCase__ : str =tax_attention_out lowerCamelCase__ : Optional[int] =tax_attention_query lowerCamelCase__ : Dict =tax_attention_value lowerCamelCase__ : List[str] =tax_pre_attention_layer_norm lowerCamelCase__ : List[Any] =tax_enc_dec_attention_key lowerCamelCase__ : Any =tax_enc_dec_attention_out lowerCamelCase__ : Any =tax_enc_dec_attention_query lowerCamelCase__ : Optional[int] =tax_enc_dec_attention_value lowerCamelCase__ : Dict =tax_cross_layer_norm if split_mlp_wi: lowerCamelCase__ : Tuple =tax_mlp_wi_a lowerCamelCase__ : int =tax_mlp_wi_a else: lowerCamelCase__ : List[Any] =tax_mlp_wi lowerCamelCase__ : Dict =tax_mlp_wo lowerCamelCase__ : Tuple =txa_mlp_layer_norm lowerCamelCase__ : Optional[Any] =flax_model_decoder_layer_block # Decoder Normalization lowerCamelCase__ : Dict =tax_model['''target''']['''decoder''']['''decoder_norm''']['''scale'''] lowerCamelCase__ : int =txa_decoder_norm # Only for layer 0: lowerCamelCase__ : Tuple =tax_model['''target''']['''decoder''']['''relpos_bias''']['''rel_embedding'''].T lowerCamelCase__ : Tuple =tax_decoder_rel_embedding # Token Embeddings lowerCamelCase__ : Union[str, Any] =tax_model['''target''']['''token_embedder''']['''embedding'''] lowerCamelCase__ : Dict =txa_token_embeddings # LM Head (only in v1.1 and LongT5 checkpoints) if "logits_dense" in tax_model["target"]["decoder"]: lowerCamelCase__ : int =tax_model['''target''']['''decoder''']['''logits_dense''']['''kernel'''] flax_model.save_pretrained(__lowerCamelCase ) print('''T5X Model was sucessfully converted!''' ) if __name__ == "__main__": _lowercase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--t5x_checkpoint_path", default=None, type=str, required=True, help="Path the T5X checkpoint." ) parser.add_argument("--config_name", default=None, type=str, required=True, help="Config name of LongT5/T5 model.") parser.add_argument( "--flax_dump_folder_path", default=None, type=str, required=True, help="Path to the output FLAX model." ) _lowercase : List[Any] = parser.parse_args() convert_tax_checkpoint_to_flax(args.tax_checkpoint_path, args.config_name, args.flax_dump_folder_path)
625
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase : int = logging.get_logger(__name__) _lowercase : List[str] = { "naver-clova-ix/donut-base": "https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json", # See all Donut models at https://huggingface.co/models?filter=donut-swin } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 'donut-swin' _a = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any], lowerCamelCase : int=224, lowerCamelCase : Any=4, lowerCamelCase : Optional[Any]=3, lowerCamelCase : List[Any]=96, lowerCamelCase : Dict=[2, 2, 6, 2], lowerCamelCase : Union[str, Any]=[3, 6, 12, 24], lowerCamelCase : int=7, lowerCamelCase : str=4.0, lowerCamelCase : List[str]=True, lowerCamelCase : Tuple=0.0, lowerCamelCase : Any=0.0, lowerCamelCase : List[Any]=0.1, lowerCamelCase : List[str]="gelu", lowerCamelCase : int=False, lowerCamelCase : Union[str, Any]=0.02, lowerCamelCase : str=1E-5, **lowerCamelCase : List[str], )-> List[Any]: super().__init__(**lowerCamelCase ) lowerCamelCase__ : List[Any] =image_size lowerCamelCase__ : Dict =patch_size lowerCamelCase__ : Tuple =num_channels lowerCamelCase__ : str =embed_dim lowerCamelCase__ : Union[str, Any] =depths lowerCamelCase__ : str =len(lowerCamelCase ) lowerCamelCase__ : Optional[Any] =num_heads lowerCamelCase__ : int =window_size lowerCamelCase__ : Tuple =mlp_ratio lowerCamelCase__ : Dict =qkv_bias lowerCamelCase__ : int =hidden_dropout_prob lowerCamelCase__ : List[str] =attention_probs_dropout_prob lowerCamelCase__ : int =drop_path_rate lowerCamelCase__ : Union[str, Any] =hidden_act lowerCamelCase__ : str =use_absolute_embeddings lowerCamelCase__ : Any =layer_norm_eps lowerCamelCase__ : Union[str, Any] =initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase__ : Optional[int] =int(embed_dim * 2 ** (len(lowerCamelCase ) - 1) )
625
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[Any], lowerCamelCase : Tuple, lowerCamelCase : List[str]=13, lowerCamelCase : List[Any]=32, lowerCamelCase : Dict=3, lowerCamelCase : int=4, lowerCamelCase : str=[10, 20, 30, 40], lowerCamelCase : Any=[2, 2, 3, 2], lowerCamelCase : int=True, lowerCamelCase : int=True, lowerCamelCase : str=37, lowerCamelCase : Optional[int]="gelu", lowerCamelCase : Optional[int]=10, lowerCamelCase : Any=0.02, lowerCamelCase : Union[str, Any]=["stage2", "stage3", "stage4"], lowerCamelCase : Optional[int]=3, lowerCamelCase : Tuple=None, )-> List[str]: lowerCamelCase__ : List[str] =parent lowerCamelCase__ : Tuple =batch_size lowerCamelCase__ : str =image_size lowerCamelCase__ : Any =num_channels lowerCamelCase__ : Tuple =num_stages lowerCamelCase__ : List[str] =hidden_sizes lowerCamelCase__ : Any =depths lowerCamelCase__ : Union[str, Any] =is_training lowerCamelCase__ : Tuple =use_labels lowerCamelCase__ : int =intermediate_size lowerCamelCase__ : Optional[int] =hidden_act lowerCamelCase__ : Dict =type_sequence_label_size lowerCamelCase__ : Tuple =initializer_range lowerCamelCase__ : Any =out_features lowerCamelCase__ : Tuple =num_labels lowerCamelCase__ : Optional[int] =scope lowerCamelCase__ : Optional[int] =num_stages def snake_case ( self : str )-> Optional[int]: lowerCamelCase__ : str =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Tuple =None if self.use_labels: lowerCamelCase__ : Union[str, Any] =ids_tensor([self.batch_size], self.type_sequence_label_size ) lowerCamelCase__ : int =self.get_config() return config, pixel_values, labels def snake_case ( self : Union[str, Any] )-> Any: return ConvNextConfig( num_channels=self.num_channels, num_stages=self.num_stages, hidden_sizes=self.hidden_sizes, depths=self.depths, is_training=self.is_training, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, out_features=self.out_features, ) def snake_case ( self : Union[str, Any] )-> Any: return UperNetConfig( backbone_config=self.get_backbone_config(), hidden_size=512, pool_scales=[1, 2, 3, 6], use_auxiliary_head=lowerCamelCase, auxiliary_loss_weight=0.4, auxiliary_in_channels=40, auxiliary_channels=256, auxiliary_num_convs=1, auxiliary_concat_input=lowerCamelCase, loss_ignore_index=255, num_labels=self.num_labels, ) def snake_case ( self : int, lowerCamelCase : str, lowerCamelCase : List[str], lowerCamelCase : List[Any] )-> Tuple: lowerCamelCase__ : List[str] =UperNetForSemanticSegmentation(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() lowerCamelCase__ : int =model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape, (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def snake_case ( self : Any )-> Tuple: lowerCamelCase__ : Dict =self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Any =config_and_inputs lowerCamelCase__ : Optional[int] ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' _a = (UperNetForSemanticSegmentation,) if is_torch_available() else () _a = {'image-segmentation': UperNetForSemanticSegmentation} if is_torch_available() else {} _a = False _a = False _a = False _a = False _a = False _a = False def snake_case ( self : Optional[int] )-> Optional[int]: lowerCamelCase__ : Optional[Any] =UperNetModelTester(self ) lowerCamelCase__ : Union[str, Any] =ConfigTester(self, config_class=lowerCamelCase, has_text_modality=lowerCamelCase, hidden_size=37 ) def snake_case ( self : Optional[int] )-> Optional[int]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case ( self : List[str] )-> Dict: return def snake_case ( self : Optional[int] )-> List[str]: lowerCamelCase__ , lowerCamelCase__ : str =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Union[str, Any] =model_class(lowerCamelCase ) lowerCamelCase__ : Tuple =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Tuple =[*signature.parameters.keys()] lowerCamelCase__ : List[Any] =['''pixel_values'''] self.assertListEqual(arg_names[:1], lowerCamelCase ) def snake_case ( self : Any )-> Union[str, Any]: lowerCamelCase__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @unittest.skip(reason='''UperNet does not use inputs_embeds''' ) def snake_case ( self : Optional[Any] )-> List[Any]: pass @unittest.skip(reason='''UperNet does not support input and output embeddings''' ) def snake_case ( self : Any )-> List[str]: pass @unittest.skip(reason='''UperNet does not have a base model''' ) def snake_case ( self : int )-> Any: pass @unittest.skip(reason='''UperNet does not have a base model''' ) def snake_case ( self : Dict )-> str: pass @require_torch_multi_gpu @unittest.skip(reason='''UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def snake_case ( self : List[Any] )-> List[str]: pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def snake_case ( self : Tuple )-> str: pass def snake_case ( self : Optional[int] )-> List[str]: def check_hidden_states_output(lowerCamelCase : Dict, lowerCamelCase : int, lowerCamelCase : List[str] ): lowerCamelCase__ : Union[str, Any] =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Optional[Any] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : Optional[Any] =outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCamelCase__ : List[str] =self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ), expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ), [self.model_tester.image_size // 4, self.model_tester.image_size // 4], ) lowerCamelCase__ , lowerCamelCase__ : Dict =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] =True check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] =True check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : Any )-> List[Any]: lowerCamelCase__ , lowerCamelCase__ : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : str =_config_zero_init(lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =_config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] =model_class(config=lowerCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item(), [0.0, 1.0], msg=F'''Parameter {name} of model {model_class} seems not properly initialized''', ) @unittest.skip(reason='''UperNet does not have tied weights''' ) def snake_case ( self : Any )-> str: pass @slow def snake_case ( self : int )-> Union[str, Any]: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : str =UperNetForSemanticSegmentation.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : Optional[int] =hf_hub_download( repo_id='''hf-internal-testing/fixtures_ade20k''' , repo_type='''dataset''' , filename='''ADE_val_00000001.jpg''' ) lowerCamelCase__ : List[str] =Image.open(__lowerCamelCase ).convert('''RGB''' ) return image @require_torch @require_vision @slow class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def snake_case ( self : str )-> Union[str, Any]: lowerCamelCase__ : List[Any] =AutoImageProcessor.from_pretrained('''openmmlab/upernet-swin-tiny''' ) lowerCamelCase__ : List[Any] =UperNetForSemanticSegmentation.from_pretrained('''openmmlab/upernet-swin-tiny''' ).to(lowerCamelCase ) lowerCamelCase__ : List[Any] =prepare_img() lowerCamelCase__ : List[Any] =processor(images=lowerCamelCase, return_tensors='''pt''' ).to(lowerCamelCase ) with torch.no_grad(): lowerCamelCase__ : List[Any] =model(**lowerCamelCase ) lowerCamelCase__ : Optional[int] =torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape, lowerCamelCase ) lowerCamelCase__ : Dict =torch.tensor( [[-7.5_958, -7.5_958, -7.4_302], [-7.5_958, -7.5_958, -7.4_302], [-7.4_797, -7.4_797, -7.3_068]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3], lowerCamelCase, atol=1E-4 ) ) def snake_case ( self : Optional[int] )-> Optional[Any]: lowerCamelCase__ : str =AutoImageProcessor.from_pretrained('''openmmlab/upernet-convnext-tiny''' ) lowerCamelCase__ : Tuple =UperNetForSemanticSegmentation.from_pretrained('''openmmlab/upernet-convnext-tiny''' ).to(lowerCamelCase ) lowerCamelCase__ : Dict =prepare_img() lowerCamelCase__ : Any =processor(images=lowerCamelCase, return_tensors='''pt''' ).to(lowerCamelCase ) with torch.no_grad(): lowerCamelCase__ : Any =model(**lowerCamelCase ) lowerCamelCase__ : Dict =torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape, lowerCamelCase ) lowerCamelCase__ : List[str] =torch.tensor( [[-8.8_110, -8.8_110, -8.6_521], [-8.8_110, -8.8_110, -8.6_521], [-8.7_746, -8.7_746, -8.6_130]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3], lowerCamelCase, atol=1E-4 ) )
625
1