code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" _A : Union[str, Any] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def __magic_name__ ( __snake_case : int ) -> Union[str, Any]: lowercase : Union[str, Any] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _A : List[str] = [None] * 10_00_00_00 _A : Optional[Any] = True _A : List[Any] = False def __magic_name__ ( __snake_case : int ) -> Tuple: if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore lowercase : List[Any] = chain(next_number(__a ) ) lowercase : Union[str, Any] = number_chain while number < 1000_0000: lowercase : Optional[Any] = number_chain number *= 10 return number_chain def __magic_name__ ( __snake_case : int = 1000_0000 ) -> Union[str, Any]: for i in range(1 , __a ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__a ) if __name__ == "__main__": import doctest doctest.testmod() print(F"{solution() = }")
202
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" from jiwer import compute_measures import datasets lowercase__ : Optional[int] = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ lowercase__ : Any = """\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. """ lowercase__ : Optional[int] = """ Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> wer = datasets.load_metric(\"wer\") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/jitsi/jiwer/'] , reference_urls=[ 'https://en.wikipedia.org/wiki/Word_error_rate', ] , ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , SCREAMING_SNAKE_CASE_ : Any=None , SCREAMING_SNAKE_CASE_ : List[Any]=None , SCREAMING_SNAKE_CASE_ : List[str]=False ): if concatenate_texts: return compute_measures(_a , _a )["wer"] else: lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Optional[Any] = 0 for prediction, reference in zip(_a , _a ): lowerCAmelCase_ : Dict = compute_measures(_a , _a ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
224
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
class UpperCAmelCase_ : '''simple docstring''' def __init__( self ): """simple docstring""" lowerCamelCase : Dict = {} def _snake_case ( self ): """simple docstring""" print(self.vertex ) for i in self.vertex: print(_a , " -> " , " -> ".join([str(_a ) for j in self.vertex[i]] ) ) def _snake_case ( self , __A , __A ): """simple docstring""" if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex lowerCamelCase : int = [to_vertex] def _snake_case ( self ): """simple docstring""" lowerCamelCase : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a , _a ) def _snake_case ( self , __A , __A ): """simple docstring""" lowerCamelCase : List[Any] = True print(_a , end=" " ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a , _a ) if __name__ == "__main__": _snake_case = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('''DFS:''') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
283
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _A = logging.getLogger(__name__) _A = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _A = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowercase_ : A__ : Optional[str] = field( default=lowercase__ , metadata={ """help""": ( """The model checkpoint for weights initialization. Leave None if you want to train a model from""" """ scratch.""" ) } , ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(lowercase__ )} , ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class lowercase_ : A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """The input training data file (a text file)."""} ) A__ : Optional[str] = field( default=lowercase__ , metadata={ """help""": ( """The input training data files (multiple files in glob format). """ """Very often splitting large files to smaller files can prevent tokenizer going out of memory""" ) } , ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """An optional input train ref data file for whole word mask in Chinese."""} , ) A__ : Optional[str] = field( default=lowercase__ , metadata={"""help""": """An optional input eval ref data file for whole word mask in Chinese."""} , ) A__ : bool = field( default=lowercase__ , metadata={"""help""": """Whether distinct lines of text in the dataset are to be handled as distinct sequences."""} , ) A__ : bool = field( default=lowercase__ , metadata={"""help""": """Train with masked-language modeling loss instead of language modeling."""} ) A__ : bool = field(default=lowercase__ , metadata={"""help""": """Whether ot not to use whole word mask."""} ) A__ : float = field( default=0.15 , metadata={"""help""": """Ratio of tokens to mask for masked language modeling loss"""} ) A__ : float = field( default=1 / 6 , metadata={ """help""": ( """Ratio of length of a span of masked tokens to surrounding context length for permutation language""" """ modeling.""" ) } , ) A__ : int = field( default=5 , metadata={"""help""": """Maximum length of a span of masked tokens for permutation language modeling."""} ) A__ : int = field( default=-1 , metadata={ """help""": ( """Optional input sequence length after tokenization.""" """The training dataset will be truncated in block of this size for training.""" """Default to the model max input length for single sentence inputs (take into account special tokens).""" ) } , ) A__ : bool = field( default=lowercase__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def lowerCamelCase__ ( a__ : DataTrainingArguments , a__ : PreTrainedTokenizer , a__ : bool = False , a__ : Optional[str] = None , ) -> Union[str, Any]: def _dataset(a__ : Any , a__ : int=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError("""You need to set world whole masking and mlm to True for Chinese Whole Word Mask""" ) return LineByLineWithRefDataset( tokenizer=__a , file_path=__a , block_size=args.block_size , ref_path=__a , ) return LineByLineTextDataset(tokenizer=__a , file_path=__a , block_size=args.block_size ) else: return TextDataset( tokenizer=__a , file_path=__a , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=__a , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(__a ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def lowerCamelCase__ ( ) -> Any: UpperCamelCase_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCamelCase_ = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( """Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file """ """or remove the --do_eval argument.""" ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , __a ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: UpperCamelCase_ = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCamelCase_ = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: UpperCamelCase_ = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.tokenizer_name: UpperCamelCase_ = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCamelCase_ = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( """You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another""" """ script, save it,and load it from here, using --tokenizer_name""" ) if model_args.model_name_or_path: UpperCamelCase_ = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=__a , cache_dir=model_args.cache_dir , ) else: logger.info("""Training new model from scratch""" ) UpperCamelCase_ = AutoModelWithLMHead.from_config(__a ) model.resize_token_embeddings(len(__a ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( """BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the""" """--mlm flag (masked language modeling).""" ) if data_args.block_size <= 0: UpperCamelCase_ = tokenizer.max_len # Our input block size will be the max possible for the model else: UpperCamelCase_ = min(data_args.block_size , tokenizer.max_len ) # Get datasets UpperCamelCase_ = ( get_dataset(__a , tokenizer=__a , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) UpperCamelCase_ = ( get_dataset(__a , tokenizer=__a , evaluate=__a , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": UpperCamelCase_ = DataCollatorForPermutationLanguageModeling( tokenizer=__a , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: UpperCamelCase_ = DataCollatorForWholeWordMask( tokenizer=__a , mlm_probability=data_args.mlm_probability ) else: UpperCamelCase_ = DataCollatorForLanguageModeling( tokenizer=__a , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer UpperCamelCase_ = Trainer( model=__a , args=__a , data_collator=__a , train_dataset=__a , eval_dataset=__a , prediction_loss_only=__a , ) # Training if training_args.do_train: UpperCamelCase_ = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=__a ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCamelCase_ = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) UpperCamelCase_ = trainer.evaluate() UpperCamelCase_ = math.exp(eval_output["""eval_loss"""] ) UpperCamelCase_ = {'perplexity': perplexity} UpperCamelCase_ = os.path.join(training_args.output_dir , """eval_results_lm.txt""" ) if trainer.is_world_master(): with open(__a , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key in sorted(result.keys() ): logger.info(""" %s = %s""" , __a , str(result[key] ) ) writer.write("""%s = %s\n""" % (key, str(result[key] )) ) results.update(__a ) return results def lowerCamelCase__ ( a__ : Optional[int] ) -> Any: main() if __name__ == "__main__": main()
122
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' import os import string import sys A =1 << 8 A ={ 'tab': ord('\t'), 'newline': ord('\r'), 'esc': 27, 'up': 65 + ARROW_KEY_FLAG, 'down': 66 + ARROW_KEY_FLAG, 'right': 67 + ARROW_KEY_FLAG, 'left': 68 + ARROW_KEY_FLAG, 'mod_int': 91, 'undefined': sys.maxsize, 'interrupt': 3, 'insert': 50, 'delete': 51, 'pg_up': 53, 'pg_down': 54, } A =KEYMAP['up'] A =KEYMAP['left'] if sys.platform == "win32": A =[] A ={ B'\xe0H': KEYMAP['up'] - ARROW_KEY_FLAG, B'\x00H': KEYMAP['up'] - ARROW_KEY_FLAG, B'\xe0P': KEYMAP['down'] - ARROW_KEY_FLAG, B'\x00P': KEYMAP['down'] - ARROW_KEY_FLAG, B'\xe0M': KEYMAP['right'] - ARROW_KEY_FLAG, B'\x00M': KEYMAP['right'] - ARROW_KEY_FLAG, B'\xe0K': KEYMAP['left'] - ARROW_KEY_FLAG, B'\x00K': KEYMAP['left'] - ARROW_KEY_FLAG, } for i in range(10): A =ord(str(i)) def snake_case_ (): if os.name == "nt": import msvcrt UpperCAmelCase = 'mbcs' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(__a ) == 0: # Read the keystroke UpperCAmelCase = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): UpperCAmelCase = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: UpperCAmelCase = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['''mod_int'''] ) ) WIN_CH_BUFFER.append(__a ) if ord(__a ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(1_2_6 ) ) UpperCAmelCase = chr(KEYMAP['''esc'''] ) except KeyError: UpperCAmelCase = cha[1] else: UpperCAmelCase = ch.decode(__a ) else: UpperCAmelCase = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty UpperCAmelCase = sys.stdin.fileno() UpperCAmelCase = termios.tcgetattr(__a ) try: tty.setraw(__a ) UpperCAmelCase = sys.stdin.read(1 ) finally: termios.tcsetattr(__a , termios.TCSADRAIN , __a ) return ch def snake_case_ (): UpperCAmelCase = get_raw_chars() if ord(__a ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(__a ) == KEYMAP["esc"]: UpperCAmelCase = get_raw_chars() if ord(__a ) == KEYMAP["mod_int"]: UpperCAmelCase = get_raw_chars() if ord(__a ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(__a ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(__a ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
34
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
import pprint import requests __UpperCamelCase : Union[str, Any] = """https://zenquotes.io/api""" def a_ ( ) -> Tuple: """simple docstring""" return requests.get(API_ENDPOINT_URL + '/today' ).json() def a_ ( ) -> List[str]: """simple docstring""" return requests.get(API_ENDPOINT_URL + '/random' ).json() if __name__ == "__main__": __UpperCamelCase : Optional[int] = random_quotes() pprint.pprint(response)
307
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter a : int = 'Create a default config file for Accelerate with only a few flags set.' def lowerCAmelCase_ (lowerCAmelCase__: Tuple="no" , lowerCAmelCase__: str = default_json_config_file , lowerCAmelCase__: bool = False ): """simple docstring""" UpperCAmelCase_: List[Any] = Path(__a ) path.parent.mkdir(parents=__a , exist_ok=__a ) if path.exists(): print( F'Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.' ) return False UpperCAmelCase_: int = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( F'`mixed_precision` should be one of \'no\', \'fp16\', \'bf16\', or \'fp8\'. Received {mixed_precision}' ) UpperCAmelCase_: List[Any] = { 'compute_environment': 'LOCAL_MACHINE', 'mixed_precision': mixed_precision, } if torch.cuda.is_available(): UpperCAmelCase_: Optional[int] = torch.cuda.device_count() UpperCAmelCase_: int = num_gpus UpperCAmelCase_: Any = False if num_gpus > 1: UpperCAmelCase_: Union[str, Any] = 'MULTI_GPU' else: UpperCAmelCase_: List[str] = 'NO' elif is_xpu_available() and use_xpu: UpperCAmelCase_: Dict = torch.xpu.device_count() UpperCAmelCase_: Tuple = num_xpus UpperCAmelCase_: List[Any] = False if num_xpus > 1: UpperCAmelCase_: Union[str, Any] = 'MULTI_XPU' else: UpperCAmelCase_: Any = 'NO' elif is_npu_available(): UpperCAmelCase_: int = torch.npu.device_count() UpperCAmelCase_: Optional[Any] = num_npus UpperCAmelCase_: List[Any] = False if num_npus > 1: UpperCAmelCase_: List[Any] = 'MULTI_NPU' else: UpperCAmelCase_: List[str] = 'NO' else: UpperCAmelCase_: Dict = 0 UpperCAmelCase_: int = True UpperCAmelCase_: Optional[int] = 1 UpperCAmelCase_: Dict = 'NO' UpperCAmelCase_: Optional[int] = ClusterConfig(**__a ) config.to_json_file(__a ) return path def lowerCAmelCase_ (lowerCAmelCase__: str , lowerCAmelCase__: Optional[Any] ): """simple docstring""" UpperCAmelCase_: Any = parser.add_parser("""default""" , parents=__a , help=__a , formatter_class=__a ) parser.add_argument( """--config_file""" , default=__a , help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have """ """such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed """ """with \'huggingface\'.""" ) , dest="""save_location""" , ) parser.add_argument( """--mixed_precision""" , choices=["""no""", """fp16""", """bf16"""] , type=__a , help="""Whether or not to use mixed precision training. """ """Choose between FP16 and BF16 (bfloat16) training. """ """BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.""" , default="""no""" , ) parser.set_defaults(func=__a ) return parser def lowerCAmelCase_ (lowerCAmelCase__: List[str] ): """simple docstring""" UpperCAmelCase_: Union[str, Any] = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(F'accelerate configuration saved at {config_file}' )
147
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer _snake_case : Optional[int] = logging.get_logger(__name__) _snake_case : List[Any] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _snake_case : Union[str, Any] = { 'vocab_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt' ), 'squeezebert/squeezebert-mnli': 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt', 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli': ( 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json' ), }, } _snake_case : Tuple = { 'squeezebert/squeezebert-uncased': 512, 'squeezebert/squeezebert-mnli': 512, 'squeezebert/squeezebert-mnli-headless': 512, } _snake_case : Any = { 'squeezebert/squeezebert-uncased': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli-headless': {'do_lower_case': True}, } class _UpperCAmelCase ( lowercase__ ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = PRETRAINED_VOCAB_FILES_MAP a_ = PRETRAINED_INIT_CONFIGURATION a_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = SqueezeBertTokenizer def __init__( self : List[str] , lowerCAmelCase_ : str=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : int=True , lowerCAmelCase_ : Any="[UNK]" , lowerCAmelCase_ : str="[SEP]" , lowerCAmelCase_ : Any="[PAD]" , lowerCAmelCase_ : List[str]="[CLS]" , lowerCAmelCase_ : int="[MASK]" , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Any=None , **lowerCAmelCase_ : Tuple , ) -> Optional[Any]: super().__init__( _a , tokenizer_file=_a , do_lower_case=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , tokenize_chinese_chars=_a , strip_accents=_a , **_a , ) __lowerCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , _a ) != do_lower_case or normalizer_state.get('strip_accents' , _a ) != strip_accents or normalizer_state.get('handle_chinese_chars' , _a ) != tokenize_chinese_chars ): __lowerCAmelCase = getattr(_a , normalizer_state.pop('type' ) ) __lowerCAmelCase = do_lower_case __lowerCAmelCase = strip_accents __lowerCAmelCase = tokenize_chinese_chars __lowerCAmelCase = normalizer_class(**_a ) __lowerCAmelCase = do_lower_case def lowercase ( self : Tuple , lowerCAmelCase_ : Any , lowerCAmelCase_ : Dict=None ) -> Tuple: __lowerCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase ( self : int , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> str: __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase ( self : Tuple , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ) -> List[str]: __lowerCAmelCase = self._tokenizer.model.save(_a , name=_a ) return tuple(_a )
284
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCamelCase = { '''configuration_vision_text_dual_encoder''': ['''VisionTextDualEncoderConfig'''], '''processing_vision_text_dual_encoder''': ['''VisionTextDualEncoderProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''VisionTextDualEncoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''FlaxVisionTextDualEncoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''TFVisionTextDualEncoderModel'''] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
131
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" from __future__ import annotations from collections.abc import Callable def a__ ( SCREAMING_SNAKE_CASE : Callable[[int | float], int | float] , SCREAMING_SNAKE_CASE : int | float , SCREAMING_SNAKE_CASE : int | float , SCREAMING_SNAKE_CASE : int = 1_0_0 , ): '''simple docstring''' lowerCAmelCase : Dict = x_start lowerCAmelCase : Dict = fnc(__a ) lowerCAmelCase : Union[str, Any] = 0.0 for _ in range(__a ): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase : List[Any] = (x_end - x_start) / steps + xa lowerCAmelCase : Dict = fnc(__a ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step lowerCAmelCase : Any = xa lowerCAmelCase : Union[str, Any] = fxa return area if __name__ == "__main__": def a__ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') lowerCAmelCase__ = 10 while i <= 100_000: print(F"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
108
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def lowercase ( _snake_case : List[Any] ) ->Tuple: """simple docstring""" if ( (cp >= 0x4e00 and cp <= 0x9fff) or (cp >= 0x3400 and cp <= 0x4dbf) # or (cp >= 0x2_0000 and cp <= 0x2_a6df) # or (cp >= 0x2_a700 and cp <= 0x2_b73f) # or (cp >= 0x2_b740 and cp <= 0x2_b81f) # or (cp >= 0x2_b820 and cp <= 0x2_ceaf) # or (cp >= 0xf900 and cp <= 0xfaff) or (cp >= 0x2_f800 and cp <= 0x2_fa1f) # ): # return True return False def lowercase ( _snake_case : str ) ->Optional[int]: """simple docstring""" for char in word: __snake_case : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def lowercase ( _snake_case : List[str] ) ->List[str]: """simple docstring""" __snake_case : Dict = set() for token in tokens: __snake_case : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) __snake_case : Optional[Any] = list(__a ) return word_list def lowercase ( _snake_case : List[str] , _snake_case : set() ) ->List[str]: """simple docstring""" if not chinese_word_set: return bert_tokens __snake_case : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) __snake_case : Optional[int] = bert_tokens __snake_case : Any = 0, len(__a ) while start < end: __snake_case : Tuple = True if is_chinese(bert_word[start] ): __snake_case : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): __snake_case : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): __snake_case : Any = '##' + bert_word[j] __snake_case : Union[str, Any] = start + i __snake_case : int = False break if single_word: start += 1 return bert_word def lowercase ( _snake_case : List[str] , _snake_case : LTP , _snake_case : BertTokenizer ) ->Union[str, Any]: """simple docstring""" __snake_case : int = [] for i in range(0 , len(__a ) , 100 ): __snake_case : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 100] )[0] __snake_case : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) __snake_case : str = [] for i in range(0 , len(__a ) , 100 ): __snake_case : List[str] = bert_tokenizer(lines[i : i + 100] , add_special_tokens=__a , truncation=__a , max_length=512 ) bert_res.extend(res['''input_ids'''] ) assert len(__a ) == len(__a ) __snake_case : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): __snake_case : int = [] for id in input_ids: __snake_case : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) __snake_case : List[str] = add_sub_symbol(__a , __a ) __snake_case : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": __snake_case : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def lowercase ( _snake_case : Optional[Any] ) ->Any: """simple docstring""" with open(args.file_name , '''r''' , encoding='''utf-8''' ) as f: __snake_case : Dict = f.readlines() __snake_case : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' __snake_case : int = LTP(args.ltp ) # faster in GPU device __snake_case : Tuple = BertTokenizer.from_pretrained(args.bert ) __snake_case : int = prepare_ref(__a , __a , __a ) with open(args.save_path , '''w''' , encoding='''utf-8''' ) as f: __snake_case : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Dict = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() main(args)
102
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" from __future__ import annotations from random import choice def __magic_name__ ( __snake_case : str ) -> List[Any]: return choice(__a ) def __magic_name__ ( __snake_case : list[int] , __snake_case : int ) -> List[str]: lowercase : Dict = random_pivot(__a ) # partition based on pivot # linear time lowercase : Optional[int] = [e for e in lst if e < pivot] lowercase : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
202
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : str = logging.get_logger(__name__) lowercase__ : Optional[int] = { """Salesforce/blip-vqa-base""": """https://huggingface.co/Salesforce/blip-vqa-base/resolve/main/config.json""", """Salesforce/blip-vqa-capfit-large""": ( """https://huggingface.co/Salesforce/blip-vqa-base-capfit/resolve/main/config.json""" ), """Salesforce/blip-image-captioning-base""": ( """https://huggingface.co/Salesforce/blip-image-captioning-base/resolve/main/config.json""" ), """Salesforce/blip-image-captioning-large""": ( """https://huggingface.co/Salesforce/blip-image-captioning-large/resolve/main/config.json""" ), """Salesforce/blip-itm-base-coco""": """https://huggingface.co/Salesforce/blip-itm-base-coco/resolve/main/config.json""", """Salesforce/blip-itm-large-coco""": """https://huggingface.co/Salesforce/blip-itm-large-coco/resolve/main/config.json""", """Salesforce/blip-itm-base-flikr""": """https://huggingface.co/Salesforce/blip-itm-base-flikr/resolve/main/config.json""", """Salesforce/blip-itm-large-flikr""": ( """https://huggingface.co/Salesforce/blip-itm-large-flikr/resolve/main/config.json""" ), } class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = '''blip_text_model''' def __init__( self : Any , SCREAMING_SNAKE_CASE_ : List[Any]=3_0_5_2_4 , SCREAMING_SNAKE_CASE_ : Optional[Any]=7_6_8 , SCREAMING_SNAKE_CASE_ : Dict=7_6_8 , SCREAMING_SNAKE_CASE_ : Optional[int]=3_0_7_2 , SCREAMING_SNAKE_CASE_ : str=7_6_8 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=1_2 , SCREAMING_SNAKE_CASE_ : List[str]=8 , SCREAMING_SNAKE_CASE_ : str=5_1_2 , SCREAMING_SNAKE_CASE_ : List[Any]="gelu" , SCREAMING_SNAKE_CASE_ : Any=1E-12 , SCREAMING_SNAKE_CASE_ : Tuple=0.0 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=0.0 , SCREAMING_SNAKE_CASE_ : Tuple=0.02 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=3_0_5_2_2 , SCREAMING_SNAKE_CASE_ : str=2 , SCREAMING_SNAKE_CASE_ : List[str]=0 , SCREAMING_SNAKE_CASE_ : Tuple=1_0_2 , SCREAMING_SNAKE_CASE_ : List[Any]=True , SCREAMING_SNAKE_CASE_ : Any=True , **SCREAMING_SNAKE_CASE_ : List[Any] , ): super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , sep_token_id=_a , **_a , ) lowerCAmelCase_ : Any = vocab_size lowerCAmelCase_ : str = hidden_size lowerCAmelCase_ : Any = encoder_hidden_size lowerCAmelCase_ : Optional[int] = intermediate_size lowerCAmelCase_ : Any = projection_dim lowerCAmelCase_ : Any = hidden_dropout_prob lowerCAmelCase_ : Union[str, Any] = num_hidden_layers lowerCAmelCase_ : List[str] = num_attention_heads lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : Tuple = hidden_act lowerCAmelCase_ : Tuple = initializer_range lowerCAmelCase_ : List[str] = attention_probs_dropout_prob lowerCAmelCase_ : str = is_decoder lowerCAmelCase_ : Tuple = use_cache @classmethod def SCREAMING_SNAKE_CASE__ ( cls : int , SCREAMING_SNAKE_CASE_ : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE_ : Tuple ): cls._set_token_in_kwargs(_a ) lowerCAmelCase_ : Any = cls.get_config_dict(_a , **_a ) # get the text config dict if we are loading from BlipConfig if config_dict.get('model_type' ) == "blip": lowerCAmelCase_ : Tuple = config_dict['text_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_a , **_a ) class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = '''blip_vision_model''' def __init__( self : Tuple , SCREAMING_SNAKE_CASE_ : List[Any]=7_6_8 , SCREAMING_SNAKE_CASE_ : List[str]=3_0_7_2 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=5_1_2 , SCREAMING_SNAKE_CASE_ : Any=1_2 , SCREAMING_SNAKE_CASE_ : Tuple=1_2 , SCREAMING_SNAKE_CASE_ : Any=3_8_4 , SCREAMING_SNAKE_CASE_ : Tuple=1_6 , SCREAMING_SNAKE_CASE_ : Any="gelu" , SCREAMING_SNAKE_CASE_ : Dict=1E-5 , SCREAMING_SNAKE_CASE_ : int=0.0 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=1E-10 , **SCREAMING_SNAKE_CASE_ : List[Any] , ): super().__init__(**_a ) lowerCAmelCase_ : Tuple = hidden_size lowerCAmelCase_ : str = intermediate_size lowerCAmelCase_ : Tuple = projection_dim lowerCAmelCase_ : Union[str, Any] = num_hidden_layers lowerCAmelCase_ : Optional[Any] = num_attention_heads lowerCAmelCase_ : int = patch_size lowerCAmelCase_ : List[Any] = image_size lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : str = attention_dropout lowerCAmelCase_ : Union[str, Any] = layer_norm_eps lowerCAmelCase_ : List[str] = hidden_act @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Tuple , SCREAMING_SNAKE_CASE_ : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE_ : int ): cls._set_token_in_kwargs(_a ) lowerCAmelCase_ : List[str] = cls.get_config_dict(_a , **_a ) # get the vision config dict if we are loading from BlipConfig if config_dict.get('model_type' ) == "blip": lowerCAmelCase_ : List[Any] = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_a , **_a ) class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = '''blip''' _SCREAMING_SNAKE_CASE = True def __init__( self : Dict , SCREAMING_SNAKE_CASE_ : Optional[Any]=None , SCREAMING_SNAKE_CASE_ : List[str]=None , SCREAMING_SNAKE_CASE_ : Any=5_1_2 , SCREAMING_SNAKE_CASE_ : Optional[Any]=2.65_92 , SCREAMING_SNAKE_CASE_ : Optional[int]=2_5_6 , **SCREAMING_SNAKE_CASE_ : Optional[int] , ): super().__init__(**_a ) if text_config is None: lowerCAmelCase_ : Dict = {} logger.info('`text_config` is `None`. Initializing the `BlipTextConfig` with default values.' ) if vision_config is None: lowerCAmelCase_ : Any = {} logger.info('`vision_config` is `None`. Initializing the `BlipVisionConfig` with default values.' ) lowerCAmelCase_ : Union[str, Any] = BlipTextConfig(**_a ) lowerCAmelCase_ : Union[str, Any] = BlipVisionConfig(**_a ) lowerCAmelCase_ : int = self.vision_config.hidden_size lowerCAmelCase_ : Tuple = projection_dim lowerCAmelCase_ : int = logit_scale_init_value lowerCAmelCase_ : List[Any] = 1.0 lowerCAmelCase_ : Any = 0.02 lowerCAmelCase_ : Optional[int] = image_text_hidden_size @classmethod def SCREAMING_SNAKE_CASE__ ( cls : str , SCREAMING_SNAKE_CASE_ : BlipTextConfig , SCREAMING_SNAKE_CASE_ : BlipVisionConfig , **SCREAMING_SNAKE_CASE_ : Optional[int] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_a ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): lowerCAmelCase_ : Dict = copy.deepcopy(self.__dict__ ) lowerCAmelCase_ : int = self.text_config.to_dict() lowerCAmelCase_ : Tuple = self.vision_config.to_dict() lowerCAmelCase_ : int = self.__class__.model_type return output
224
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : Dict = len(__a ) lowerCamelCase : Optional[int] = sum(__a ) lowerCamelCase : Dict = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): lowerCamelCase : int = True for i in range(1 , s + 1 ): lowerCamelCase : Dict = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): lowerCamelCase : Tuple = dp[i][j - 1] if arr[i - 1] <= j: lowerCamelCase : List[str] = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: lowerCamelCase : str = s - 2 * j break return diff
283
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import AlignProcessor, EfficientNetImageProcessor @require_vision class lowercase_ ( unittest.TestCase ): def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = tempfile.mkdtemp() UpperCamelCase_ = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] UpperCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) UpperCamelCase_ = { 'do_resize': True, 'size': 2_0, 'do_center_crop': True, 'crop_size': 1_8, 'do_normalize': True, 'image_mean': [0.48_145_466, 0.4_578_275, 0.40_821_073], 'image_std': [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCamelCase_ = os.path.join(self.tmpdirname , _a ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(_a , _a ) def lowerCamelCase_ ( self , **__UpperCamelCase ): """simple docstring""" return BertTokenizer.from_pretrained(self.tmpdirname , **_a ) def lowerCamelCase_ ( self , **__UpperCamelCase ): """simple docstring""" return BertTokenizerFast.from_pretrained(self.tmpdirname , **_a ) def lowerCamelCase_ ( self , **__UpperCamelCase ): """simple docstring""" return EfficientNetImageProcessor.from_pretrained(self.tmpdirname , **_a ) def lowerCamelCase_ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCamelCase_ = [Image.fromarray(np.moveaxis(_a , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = self.get_rust_tokenizer() UpperCamelCase_ = self.get_image_processor() UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) processor_slow.save_pretrained(self.tmpdirname ) UpperCamelCase_ = AlignProcessor.from_pretrained(self.tmpdirname , use_fast=_a ) UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) processor_fast.save_pretrained(self.tmpdirname ) UpperCamelCase_ = AlignProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _a ) self.assertIsInstance(processor_fast.tokenizer , _a ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _a ) self.assertIsInstance(processor_fast.image_processor , _a ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = AlignProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase_ = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCamelCase_ = self.get_image_processor(do_normalize=_a , padding_value=1.0 ) UpperCamelCase_ = AlignProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=_a , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _a ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _a ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = self.get_image_processor() UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) UpperCamelCase_ = self.prepare_image_inputs() UpperCamelCase_ = image_processor(_a , return_tensors="""np""" ) UpperCamelCase_ = processor(images=_a , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = self.get_image_processor() UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) UpperCamelCase_ = 'lower newer' UpperCamelCase_ = processor(text=_a ) UpperCamelCase_ = tokenizer(_a , padding="""max_length""" , max_length=6_4 ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = self.get_image_processor() UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) UpperCamelCase_ = 'lower newer' UpperCamelCase_ = self.prepare_image_inputs() UpperCamelCase_ = processor(text=_a , images=_a ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(_a ): processor() def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = self.get_image_processor() UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) UpperCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase_ = processor.batch_decode(_a ) UpperCamelCase_ = tokenizer.batch_decode(_a ) self.assertListEqual(_a , _a ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = self.get_image_processor() UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = AlignProcessor(tokenizer=_a , image_processor=_a ) UpperCamelCase_ = 'lower newer' UpperCamelCase_ = self.prepare_image_inputs() UpperCamelCase_ = processor(text=_a , images=_a ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
122
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _a ( lowercase__ ): __a : Dict = ['''image_processor''', '''tokenizer'''] __a : Optional[int] = '''ViTImageProcessor''' __a : List[str] = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self : Dict , lowercase : List[str]=None , lowercase : List[Any]=None , **lowercase : int ): '''simple docstring''' UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _a , ) UpperCAmelCase = kwargs.pop('''feature_extractor''' ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_a , _a ) def __call__( self : str , lowercase : Tuple=None , lowercase : str=None , lowercase : Tuple=None , lowercase : int=None , **lowercase : Any ): '''simple docstring''' if text is None and visual_prompt is None and images is None: raise ValueError('''You have to specify either text, visual prompt or images.''' ) if text is not None and visual_prompt is not None: raise ValueError('''You have to specify exactly one type of prompt. Either text or visual prompt.''' ) if text is not None: UpperCAmelCase = self.tokenizer(_a , return_tensors=_a , **_a ) if visual_prompt is not None: UpperCAmelCase = self.image_processor(_a , return_tensors=_a , **_a ) if images is not None: UpperCAmelCase = self.image_processor(_a , return_tensors=_a , **_a ) if visual_prompt is not None and images is not None: UpperCAmelCase = { 'pixel_values': image_features.pixel_values, 'conditional_pixel_values': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: UpperCAmelCase = { 'conditional_pixel_values': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**_a ) , tensor_type=_a ) def A ( self : List[Any] , *lowercase : Dict , **lowercase : Any ): '''simple docstring''' return self.tokenizer.batch_decode(*_a , **_a ) def A ( self : str , *lowercase : Optional[int] , **lowercase : List[Any] ): '''simple docstring''' return self.tokenizer.decode(*_a , **_a ) @property def A ( self : Optional[int] ): '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _a , ) return self.image_processor_class @property def A ( self : Dict ): '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _a , ) return self.image_processor
34
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case__ = { 'repo_name': ['test_repo1', 'test_repo2', 'test_repo3'], 'path': ['test_1.py', 'test_2.py', 'unit_test.py'], 'content': ['a ' * 20, 'a ' * 30, 'b ' * 7], } snake_case__ = Dataset.from_dict(__a ) return dataset class __SCREAMING_SNAKE_CASE( lowercase__ ): def lowerCAmelCase_ ( self: int ) -> List[str]: snake_case__ = get_dataset() snake_case__ = make_duplicate_clusters(_a , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def lowerCAmelCase_ ( self: List[str] ) -> Optional[int]: snake_case__ = get_dataset() snake_case__ = deduplicate_dataset(_a ) self.assertEqual(len(_a ) , 2 ) print(_a ) self.assertEqual(duplicate_clusters[0][0]['copies'] , 2 ) self.assertEqual(duplicate_clusters[0][0]['is_extreme'] , _a )
307
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
from collections import deque class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: UpperCAmelCase_: Any = process_name # process name UpperCAmelCase_: Union[str, Any] = arrival_time # arrival time of the process # completion time of finished process or last interrupted time UpperCAmelCase_: Union[str, Any] = arrival_time UpperCAmelCase_: Union[str, Any] = burst_time # remaining burst time UpperCAmelCase_: Optional[Any] = 0 # total time of the process wait in ready queue UpperCAmelCase_: Tuple = 0 # time from arrival time to completion time class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, ) -> Union[str, Any]: UpperCAmelCase_: Dict = number_of_queues # time slice of queues that round robin algorithm applied UpperCAmelCase_: List[Any] = time_slices # unfinished process is in this ready_queue UpperCAmelCase_: Dict = queue # current time UpperCAmelCase_: Any = current_time # finished process is in this sequence queue UpperCAmelCase_: deque[Process] = deque() def __snake_case (self ) -> Union[str, Any]: UpperCAmelCase_: Optional[Any] = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> int: UpperCAmelCase_: Union[str, Any] = [] for i in range(len(_a ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> int: UpperCAmelCase_: Optional[Any] = [] for i in range(len(_a ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCAmelCase_: Optional[Any] = [] for i in range(len(_a ) ): completion_times.append(queue[i].stop_time ) return completion_times def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Dict: return [q.burst_time for q in queue] def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: process.waiting_time += self.current_time - process.stop_time return process.waiting_time def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> int: UpperCAmelCase_: deque[Process] = deque() # sequence deque of finished process while len(_a ) != 0: UpperCAmelCase_: List[str] = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(_a ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 UpperCAmelCase_: Dict = 0 # set the process's turnaround time because it is finished UpperCAmelCase_: List[Any] = self.current_time - cp.arrival_time # set the completion time UpperCAmelCase_: int = self.current_time # add the process to queue that has finished queue finished.append(_a ) self.finish_queue.extend(_a ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Any: UpperCAmelCase_: deque[Process] = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(_a ) ): UpperCAmelCase_: Optional[int] = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(_a ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time UpperCAmelCase_: str = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(_a ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished UpperCAmelCase_: Tuple = 0 # set the finish time UpperCAmelCase_: Optional[int] = self.current_time # update the process' turnaround time because it is finished UpperCAmelCase_: int = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(_a ) self.finish_queue.extend(_a ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def __snake_case (self ) -> Dict: for i in range(self.number_of_queues - 1 ): UpperCAmelCase_: Dict = self.round_robin( self.ready_queue, self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest a : List[Any] = Process('P1', 0, 53) a : Tuple = Process('P2', 0, 17) a : Union[str, Any] = Process('P3', 0, 68) a : List[str] = Process('P4', 0, 24) a : str = 3 a : Union[str, Any] = [17, 25] a : int = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={'queue': deque([Pa, Pa, Pa, Pa])}) a : Union[str, Any] = Process('P1', 0, 53) a : Any = Process('P2', 0, 17) a : Optional[int] = Process('P3', 0, 68) a : Dict = Process('P4', 0, 24) a : Any = 3 a : Optional[int] = [17, 25] a : List[str] = deque([Pa, Pa, Pa, Pa]) a : List[Any] = MLFQ(number_of_queues, time_slices, queue, 0) a : Optional[Any] = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F'''waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print completion times of processes(P1, P2, P3, P4) print( F'''completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print total turnaround times of processes(P1, P2, P3, P4) print( F'''turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print sequence of finished processes print( F'''sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}''' )
147
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _snake_case : Optional[int] = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. _snake_case : List[str] = direct_transformers_import(PATH_TO_TRANSFORMERS) _snake_case : str = transformers.models.auto.configuration_auto.CONFIG_MAPPING _snake_case : str = { # used to compute the property `self.chunk_length` 'EncodecConfig': ['overlap'], # used as `self.bert_model = BertModel(config, ...)` 'DPRConfig': True, # not used in modeling files, but it's an important information 'FSMTConfig': ['langs'], # used internally in the configuration class file 'GPTNeoConfig': ['attention_types'], # used internally in the configuration class file 'EsmConfig': ['is_folding_model'], # used during training (despite we don't have training script for these models yet) 'Mask2FormerConfig': ['ignore_value'], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) 'OneFormerConfig': ['ignore_value', 'norm'], # used during preprocessing and collation, see `collating_graphormer.py` 'GraphormerConfig': ['spatial_pos_max'], # used internally in the configuration class file 'T5Config': ['feed_forward_proj'], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally 'MT5Config': ['feed_forward_proj', 'tokenizer_class'], 'UMT5Config': ['feed_forward_proj', 'tokenizer_class'], # used internally in the configuration class file 'LongT5Config': ['feed_forward_proj'], # used internally in the configuration class file 'SwitchTransformersConfig': ['feed_forward_proj'], # having default values other than `1e-5` - we can't fix them without breaking 'BioGptConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'GLPNConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'SegformerConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'CvtConfig': ['layer_norm_eps'], # having default values other than `1e-5` - we can't fix them without breaking 'PerceiverConfig': ['layer_norm_eps'], # used internally to calculate the feature size 'InformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate the feature size 'TimeSeriesTransformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate the feature size 'AutoformerConfig': ['num_static_real_features', 'num_time_features'], # used internally to calculate `mlp_dim` 'SamVisionConfig': ['mlp_ratio'], # For (head) training, but so far not implemented 'ClapAudioConfig': ['num_classes'], # Not used, but providing useful information to users 'SpeechT5HifiGanConfig': ['sampling_rate'], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { 'CLIPSegConfig': True, 'DeformableDetrConfig': True, 'DetaConfig': True, 'DinatConfig': True, 'DonutSwinConfig': True, 'EfficientFormerConfig': True, 'FSMTConfig': True, 'JukeboxConfig': True, 'LayoutLMv2Config': True, 'MaskFormerSwinConfig': True, 'MT5Config': True, 'NatConfig': True, 'OneFormerConfig': True, 'PerceiverConfig': True, 'RagConfig': True, 'SpeechT5Config': True, 'SwinConfig': True, 'Swin2SRConfig': True, 'Swinv2Config': True, 'SwitchTransformersConfig': True, 'TableTransformerConfig': True, 'TapasConfig': True, 'TransfoXLConfig': True, 'UniSpeechConfig': True, 'UniSpeechSatConfig': True, 'WavLMConfig': True, 'WhisperConfig': True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) 'JukeboxPriorConfig': True, # TODO: @Younes (for `is_decoder`) 'Pix2StructTextConfig': True, } ) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Any, lowerCAmelCase_ : int, lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( F"""config.{attribute}""" in modeling_source or F"""getattr(config, \"{attribute}\"""" in modeling_source or F"""getattr(self.config, \"{attribute}\"""" in modeling_source ): __lowerCAmelCase = True # Deal with multi-line cases elif ( re.search( RF"""getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*\"{attribute}\"""", __a, ) is not None ): __lowerCAmelCase = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: __lowerCAmelCase = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files __lowerCAmelCase = [ 'bos_index', 'eos_index', 'pad_index', 'unk_index', 'mask_index', 'image_size', 'use_cache', 'out_features', 'out_indices', ] __lowerCAmelCase = ['encoder_no_repeat_ngram_size'] # Special cases to be allowed __lowerCAmelCase = True if not attribute_used: __lowerCAmelCase = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: __lowerCAmelCase = True elif attribute in ["tie_word_embeddings"] and default_value is False: __lowerCAmelCase = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: __lowerCAmelCase = True elif attribute.endswith('_token_id' ): __lowerCAmelCase = True # configuration class specific cases if not case_allowed: __lowerCAmelCase = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__, [] ) __lowerCAmelCase = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def a_ ( lowerCAmelCase_ : str ): __lowerCAmelCase = dict(inspect.signature(config_class.__init__ ).parameters ) __lowerCAmelCase = [x for x in list(signature.keys() ) if x not in ['self', 'kwargs']] __lowerCAmelCase = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass __lowerCAmelCase = {} if len(config_class.attribute_map ) > 0: __lowerCAmelCase = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files __lowerCAmelCase = inspect.getsourcefile(__a ) __lowerCAmelCase = os.path.dirname(__a ) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. __lowerCAmelCase = [os.path.join(__a, __a ) for fn in os.listdir(__a ) if fn.startswith('modeling_' )] # Get the source code strings __lowerCAmelCase = [] for path in modeling_paths: if os.path.isfile(__a ): with open(__a ) as fp: modeling_sources.append(fp.read() ) __lowerCAmelCase = [] for config_param, default_value in zip(__a, __a ): # `attributes` here is all the variant names for `config_param` __lowerCAmelCase = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param] ) if not check_attribute_being_used(__a, __a, __a, __a ): unused_attributes.append(attributes[0] ) return sorted(__a ) def a_ ( ): __lowerCAmelCase = {} for _config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) __lowerCAmelCase = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class ), lambda lowerCAmelCase_ : inspect.isclass(__a ) and issubclass(__a, __a ) and inspect.getmodule(__a ) == inspect.getmodule(_config_class ), ) ] for config_class in config_classes_in_module: __lowerCAmelCase = check_config_attributes_being_used(__a ) if len(__a ) > 0: __lowerCAmelCase = unused_attributes if len(__a ) > 0: __lowerCAmelCase = 'The following configuration classes contain unused attributes in the corresponding modeling files:\n' for name, attributes in configs_with_unused_attributes.items(): error += F"""{name}: {attributes}\n""" raise ValueError(__a ) if __name__ == "__main__": check_config_attributes()
284
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase = 5_0000 lowerCamelCase = 5000 lowerCamelCase, lowerCamelCase = os.path.split(__file__) lowerCamelCase = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCamelCase_ ( _a , _a ): """simple docstring""" for i in range(__a ): lowerCAmelCase__ : Optional[int] = dataset[i] @get_duration def lowerCamelCase_ ( _a , _a , _a ): """simple docstring""" for i in range(0 , len(__a ) , __a ): lowerCAmelCase__ : Tuple = dataset[i : i + batch_size] @get_duration def lowerCamelCase_ ( _a , _a , _a ): """simple docstring""" with dataset.formatted_as(type=__a ): for i in range(__a ): lowerCAmelCase__ : Tuple = dataset[i] @get_duration def lowerCamelCase_ ( _a , _a , _a , _a ): """simple docstring""" with dataset.formatted_as(type=__a ): for i in range(0 , __a , __a ): lowerCAmelCase__ : List[str] = dataset[i : i + batch_size] def lowerCamelCase_ ( ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = {'num examples': SPEED_TEST_N_EXAMPLES} lowerCAmelCase__ : Optional[int] = [ (read, {'length': SMALL_TEST}), (read, {'length': SPEED_TEST_N_EXAMPLES}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 10}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 100}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 1_000}), (read_formatted, {'type': 'numpy', 'length': SMALL_TEST}), (read_formatted, {'type': 'pandas', 'length': SMALL_TEST}), (read_formatted, {'type': 'torch', 'length': SMALL_TEST}), (read_formatted, {'type': 'tensorflow', 'length': SMALL_TEST}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 10}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 1_000}), ] lowerCAmelCase__ : List[Any] = [ (read, {'length': SMALL_TEST}), (read, {'length': SPEED_TEST_N_EXAMPLES}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 10}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 100}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 1_000}), (read_formatted, {'type': 'numpy', 'length': SMALL_TEST}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 10}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print('''generating dataset''' ) lowerCAmelCase__ : str = datasets.Features( {'''list''': datasets.Sequence(datasets.Value('''float32''' ) ), '''numbers''': datasets.Value('''float32''' )} ) lowerCAmelCase__ : int = generate_example_dataset( os.path.join(__a , '''dataset.arrow''' ) , __a , num_examples=__a , seq_shapes={'''list''': (100,)} , ) print('''first set of iterations''' ) for func, kwargs in functions: print(func.__name__ , str(__a ) ) lowerCAmelCase__ : List[str] = func(__a , **__a ) print('''shuffling dataset''' ) lowerCAmelCase__ : Dict = dataset.shuffle() print('''Second set of iterations (after shuffling''' ) for func, kwargs in functions_shuffled: print('''shuffled ''' , func.__name__ , str(__a ) ) lowerCAmelCase__ : Dict = func( __a , **__a ) with open(__a , '''wb''' ) as f: f.write(json.dumps(__a ).encode('''utf-8''' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
131
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self , snake_case__ , ): """simple docstring""" lowerCAmelCase : List[str] = parent lowerCAmelCase : Tuple = 13 lowerCAmelCase : Tuple = 7 lowerCAmelCase : str = 30 lowerCAmelCase : Optional[Any] = self.seq_length + self.mem_len lowerCAmelCase : Dict = 15 lowerCAmelCase : Any = True lowerCAmelCase : int = True lowerCAmelCase : List[Any] = 99 lowerCAmelCase : Tuple = [10, 50, 80] lowerCAmelCase : Optional[Any] = 32 lowerCAmelCase : str = 32 lowerCAmelCase : str = 4 lowerCAmelCase : Optional[Any] = 8 lowerCAmelCase : Tuple = 128 lowerCAmelCase : Optional[int] = 2 lowerCAmelCase : Union[str, Any] = 2 lowerCAmelCase : List[str] = None lowerCAmelCase : Tuple = 1 lowerCAmelCase : int = 0 lowerCAmelCase : Tuple = 3 lowerCAmelCase : str = self.vocab_size - 1 lowerCAmelCase : str = 0.01 def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase : int = None if self.use_labels: lowerCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase : List[Any] = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def lowercase__ ( self ): """simple docstring""" random.seed(self.seed ) tf.random.set_seed(self.seed ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : str = TFTransfoXLModel(_a ) lowerCAmelCase : str = model(_a ).to_tuple() lowerCAmelCase : Optional[int] = {'input_ids': input_ids_a, 'mems': mems_a} lowerCAmelCase : Dict = model(_a ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : Tuple = TFTransfoXLLMHeadModel(_a ) lowerCAmelCase : Any = model(_a ).to_tuple() lowerCAmelCase : Optional[int] = {'input_ids': input_ids_a, 'labels': lm_labels} lowerCAmelCase : int = model(_a ).to_tuple() lowerCAmelCase : Any = model([input_ids_a, mems_a] ).to_tuple() lowerCAmelCase : List[Any] = {'input_ids': input_ids_a, 'mems': mems_a, 'labels': lm_labels} lowerCAmelCase : List[Any] = model(_a ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : int = TFTransfoXLForSequenceClassification(_a ) lowerCAmelCase : Any = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() (lowerCAmelCase) : Any = config_and_inputs lowerCAmelCase : Union[str, Any] = {'input_ids': input_ids_a} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" a : Any =( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) a : Optional[Any] =() if is_tf_available() else () a : Union[str, Any] =( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented a : Optional[Any] =False a : Optional[Any] =False a : Union[str, Any] =False a : int =False def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[Any] = TFTransfoXLModelTester(self ) lowerCAmelCase : List[str] = ConfigTester(self , config_class=_a , d_embed=37 ) def lowercase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def lowercase__ ( self ): """simple docstring""" self.model_tester.set_seed() lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*_a ) def lowercase__ ( self ): """simple docstring""" self.model_tester.set_seed() lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*_a ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*_a ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase : str = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: lowerCAmelCase : Optional[Any] = model_class(_a ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: lowerCAmelCase : Tuple = model.get_output_embeddings() assert isinstance(_a , tf.keras.layers.Layer ) lowerCAmelCase : Tuple = model.get_bias() assert name is None else: lowerCAmelCase : Tuple = model.get_output_embeddings() assert x is None lowerCAmelCase : List[str] = model.get_bias() assert name is None def lowercase__ ( self ): """simple docstring""" pass @slow def lowercase__ ( self ): """simple docstring""" for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase : Optional[int] = TFTransfoXLModel.from_pretrained(_a ) self.assertIsNotNone(_a ) @unittest.skip(reason="This model doesn\'t play well with fit() due to not returning a single loss." ) def lowercase__ ( self ): """simple docstring""" pass @require_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" @unittest.skip("Skip test until #12651 is resolved." ) @slow def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[str] = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103" ) # fmt: off lowerCAmelCase : str = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off lowerCAmelCase : Any = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> lowerCAmelCase : List[Any] = model.generate(_a , max_length=200 , do_sample=_a ) self.assertListEqual(output_ids[0].numpy().tolist() , _a )
108
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @parameterized.expand([(None,), ('''foo.json''',)] ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' __snake_case : List[str] = GenerationConfig( do_sample=_a , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_a , config_name=_a ) __snake_case : Tuple = GenerationConfig.from_pretrained(_a , config_name=_a ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample , _a ) self.assertEqual(loaded_config.temperature , 0.7 ) self.assertEqual(loaded_config.length_penalty , 1.0 ) self.assertEqual(loaded_config.bad_words_ids , [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k , 50 ) self.assertEqual(loaded_config.max_length , 20 ) self.assertEqual(loaded_config.max_time , _a ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = AutoConfig.from_pretrained('''gpt2''' ) __snake_case : Tuple = GenerationConfig.from_model_config(_a ) __snake_case : List[Any] = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(_a , _a ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id , default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id , model_config.eos_token_id ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = GenerationConfig() __snake_case : Dict = { 'max_new_tokens': 10_24, 'foo': 'bar', } __snake_case : Optional[int] = copy.deepcopy(_a ) __snake_case : List[str] = generation_config.update(**_a ) # update_kwargs was not modified (no side effects) self.assertEqual(_a , _a ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens , 10_24 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(_a , {'''foo''': '''bar'''} ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = GenerationConfig() __snake_case : Optional[Any] = 'bar' with tempfile.TemporaryDirectory('''test-generation-config''' ) as tmp_dir: generation_config.save_pretrained(_a ) __snake_case : Optional[int] = GenerationConfig.from_pretrained(_a ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo , '''bar''' ) __snake_case : List[str] = GenerationConfig.from_model_config(_a ) assert not hasattr(_a , '''foo''' ) # no new kwargs should be initialized if from config def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = GenerationConfig() self.assertEqual(default_config.temperature , 1.0 ) self.assertEqual(default_config.do_sample , _a ) self.assertEqual(default_config.num_beams , 1 ) __snake_case : List[Any] = GenerationConfig( do_sample=_a , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) self.assertEqual(config.temperature , 0.7 ) self.assertEqual(config.do_sample , _a ) self.assertEqual(config.num_beams , 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_a ) __snake_case : Optional[Any] = GenerationConfig.from_pretrained(_a , temperature=1.0 ) self.assertEqual(loaded_config.temperature , 1.0 ) self.assertEqual(loaded_config.do_sample , _a ) self.assertEqual(loaded_config.num_beams , 1 ) # default value @is_staging_test class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @classmethod def SCREAMING_SNAKE_CASE (cls ): '''simple docstring''' __snake_case : List[str] = TOKEN HfFolder.save_token(_a ) @classmethod def SCREAMING_SNAKE_CASE (cls ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='''test-generation-config''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-generation-config-org''' ) except HTTPError: pass def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = GenerationConfig( do_sample=_a , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub('''test-generation-config''' , use_auth_token=self._token ) __snake_case : Optional[Any] = GenerationConfig.from_pretrained(f"""{USER}/test-generation-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_a , getattr(_a , _a ) ) # Reset repo delete_repo(token=self._token , repo_id='''test-generation-config''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( _a , repo_id='''test-generation-config''' , push_to_hub=_a , use_auth_token=self._token ) __snake_case : Optional[int] = GenerationConfig.from_pretrained(f"""{USER}/test-generation-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_a , getattr(_a , _a ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = GenerationConfig( do_sample=_a , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub('''valid_org/test-generation-config-org''' , use_auth_token=self._token ) __snake_case : Union[str, Any] = GenerationConfig.from_pretrained('''valid_org/test-generation-config-org''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_a , getattr(_a , _a ) ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-generation-config-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( _a , repo_id='''valid_org/test-generation-config-org''' , push_to_hub=_a , use_auth_token=self._token ) __snake_case : Tuple = GenerationConfig.from_pretrained('''valid_org/test-generation-config-org''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_a , getattr(_a , _a ) )
102
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
"""simple docstring""" import argparse import collections import os import re import tempfile import pandas as pd from datasets import Dataset from huggingface_hub import hf_hub_download, upload_folder from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/update_metadata.py _A : List[Any] = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. _A : int = direct_transformers_import(TRANSFORMERS_PATH) # Regexes that match TF/Flax/PT model names. _A : List[str] = re.compile(r"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") _A : List[Any] = re.compile(r"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. _A : Optional[int] = re.compile(r"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Fill this with tuples (pipeline_tag, model_mapping, auto_model) _A : List[str] = [ ("""pretraining""", """MODEL_FOR_PRETRAINING_MAPPING_NAMES""", """AutoModelForPreTraining"""), ("""feature-extraction""", """MODEL_MAPPING_NAMES""", """AutoModel"""), ("""audio-classification""", """MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForAudioClassification"""), ("""text-generation""", """MODEL_FOR_CAUSAL_LM_MAPPING_NAMES""", """AutoModelForCausalLM"""), ("""automatic-speech-recognition""", """MODEL_FOR_CTC_MAPPING_NAMES""", """AutoModelForCTC"""), ("""image-classification""", """MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForImageClassification"""), ("""image-segmentation""", """MODEL_FOR_IMAGE_SEGMENTATION_MAPPING_NAMES""", """AutoModelForImageSegmentation"""), ("""fill-mask""", """MODEL_FOR_MASKED_LM_MAPPING_NAMES""", """AutoModelForMaskedLM"""), ("""object-detection""", """MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES""", """AutoModelForObjectDetection"""), ( """zero-shot-object-detection""", """MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING_NAMES""", """AutoModelForZeroShotObjectDetection""", ), ("""question-answering""", """MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForQuestionAnswering"""), ("""text2text-generation""", """MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES""", """AutoModelForSeq2SeqLM"""), ("""text-classification""", """MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForSequenceClassification"""), ("""automatic-speech-recognition""", """MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES""", """AutoModelForSpeechSeq2Seq"""), ( """table-question-answering""", """MODEL_FOR_TABLE_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForTableQuestionAnswering""", ), ("""token-classification""", """MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForTokenClassification"""), ("""multiple-choice""", """MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES""", """AutoModelForMultipleChoice"""), ( """next-sentence-prediction""", """MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES""", """AutoModelForNextSentencePrediction""", ), ( """audio-frame-classification""", """MODEL_FOR_AUDIO_FRAME_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForAudioFrameClassification""", ), ("""audio-xvector""", """MODEL_FOR_AUDIO_XVECTOR_MAPPING_NAMES""", """AutoModelForAudioXVector"""), ( """document-question-answering""", """MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForDocumentQuestionAnswering""", ), ( """visual-question-answering""", """MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForVisualQuestionAnswering""", ), ("""image-to-text""", """MODEL_FOR_FOR_VISION_2_SEQ_MAPPING_NAMES""", """AutoModelForVision2Seq"""), ( """zero-shot-image-classification""", """MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForZeroShotImageClassification""", ), ("""depth-estimation""", """MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES""", """AutoModelForDepthEstimation"""), ("""video-classification""", """MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForVideoClassification"""), ("""mask-generation""", """MODEL_FOR_MASK_GENERATION_MAPPING_NAMES""", """AutoModelForMaskGeneration"""), ] def __magic_name__ ( __snake_case : Union[str, Any] ) -> int: lowercase : Optional[Any] = re.finditer(".+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)" , __a ) return [m.group(0 ) for m in matches] def __magic_name__ ( ) -> Optional[int]: lowercase : Tuple = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES lowercase : Optional[Any] = { config.replace("Config" , "" ): model_type for model_type, config in config_maping_names.items() } # Dictionaries flagging if each model prefix has a backend in PT/TF/Flax. lowercase : int = collections.defaultdict(__a ) lowercase : Tuple = collections.defaultdict(__a ) lowercase : Any = collections.defaultdict(__a ) # Let's lookup through all transformers object (once) and find if models are supported by a given backend. for attr_name in dir(__a ): lowercase : List[str] = None if _re_tf_models.match(__a ) is not None: lowercase : Tuple = tf_models lowercase : str = _re_tf_models.match(__a ).groups()[0] elif _re_flax_models.match(__a ) is not None: lowercase : Optional[Any] = flax_models lowercase : Union[str, Any] = _re_flax_models.match(__a ).groups()[0] elif _re_pt_models.match(__a ) is not None: lowercase : int = pt_models lowercase : Any = _re_pt_models.match(__a ).groups()[0] if lookup_dict is not None: while len(__a ) > 0: if attr_name in model_prefix_to_model_type: lowercase : int = True break # Try again after removing the last word in the name lowercase : Optional[int] = ''.join(camel_case_split(__a )[:-1] ) lowercase : Tuple = set(list(pt_models.keys() ) + list(tf_models.keys() ) + list(flax_models.keys() ) ) lowercase : Union[str, Any] = list(__a ) all_models.sort() lowercase : str = {'model_type': all_models} lowercase : Tuple = [pt_models[t] for t in all_models] lowercase : Tuple = [tf_models[t] for t in all_models] lowercase : Dict = [flax_models[t] for t in all_models] # Now let's use the auto-mapping names to make sure lowercase : List[str] = {} for t in all_models: if t in transformers_module.models.auto.processing_auto.PROCESSOR_MAPPING_NAMES: lowercase : Union[str, Any] = 'AutoProcessor' elif t in transformers_module.models.auto.tokenization_auto.TOKENIZER_MAPPING_NAMES: lowercase : Dict = 'AutoTokenizer' elif t in transformers_module.models.auto.feature_extraction_auto.FEATURE_EXTRACTOR_MAPPING_NAMES: lowercase : Union[str, Any] = 'AutoFeatureExtractor' else: # Default to AutoTokenizer if a model has nothing, for backward compatibility. lowercase : Optional[Any] = 'AutoTokenizer' lowercase : Optional[Any] = [processors[t] for t in all_models] return pd.DataFrame(__a ) def __magic_name__ ( __snake_case : Optional[Any] ) -> Optional[Any]: lowercase : List[Any] = [ transformers_module.models.auto.modeling_auto, transformers_module.models.auto.modeling_tf_auto, transformers_module.models.auto.modeling_flax_auto, ] for pipeline_tag, model_mapping, auto_class in PIPELINE_TAGS_AND_AUTO_MODELS: lowercase : Dict = [model_mapping, f"""TF_{model_mapping}""", f"""FLAX_{model_mapping}"""] lowercase : str = [auto_class, f"""TF_{auto_class}""", f"""Flax_{auto_class}"""] # Loop through all three frameworks for module, cls, mapping in zip(__a , __a , __a ): # The type of pipeline may not exist in this framework if not hasattr(__a , __a ): continue # First extract all model_names lowercase : int = [] for name in getattr(__a , __a ).values(): if isinstance(__a , __a ): model_names.append(__a ) else: model_names.extend(list(__a ) ) # Add pipeline tag and auto model class for those models table.update({model_name: (pipeline_tag, cls) for model_name in model_names} ) return table def __magic_name__ ( __snake_case : str , __snake_case : Any ) -> Optional[Any]: lowercase : List[str] = get_frameworks_table() lowercase : Dict = Dataset.from_pandas(__a ) lowercase : List[str] = hf_hub_download( "huggingface/transformers-metadata" , "pipeline_tags.json" , repo_type="dataset" , token=__a ) lowercase : List[Any] = Dataset.from_json(__a ) lowercase : Optional[Any] = { tags_dataset[i]['model_class']: (tags_dataset[i]['pipeline_tag'], tags_dataset[i]['auto_class']) for i in range(len(__a ) ) } lowercase : Union[str, Any] = update_pipeline_and_auto_class_table(__a ) # Sort the model classes to avoid some nondeterministic updates to create false update commits. lowercase : List[str] = sorted(table.keys() ) lowercase : Any = pd.DataFrame( { "model_class": model_classes, "pipeline_tag": [table[m][0] for m in model_classes], "auto_class": [table[m][1] for m in model_classes], } ) lowercase : List[str] = Dataset.from_pandas(__a ) with tempfile.TemporaryDirectory() as tmp_dir: frameworks_dataset.to_json(os.path.join(__a , "frameworks.json" ) ) tags_dataset.to_json(os.path.join(__a , "pipeline_tags.json" ) ) if commit_sha is not None: lowercase : str = ( f"""Update with commit {commit_sha}\n\nSee: """ f"""https://github.com/huggingface/transformers/commit/{commit_sha}""" ) else: lowercase : Optional[Any] = 'Update' upload_folder( repo_id="huggingface/transformers-metadata" , folder_path=__a , repo_type="dataset" , token=__a , commit_message=__a , ) def __magic_name__ ( ) -> Union[str, Any]: lowercase : Tuple = {tag: cls for tag, _, cls in PIPELINE_TAGS_AND_AUTO_MODELS} lowercase : Optional[Any] = transformers_module.pipelines.SUPPORTED_TASKS lowercase : Optional[int] = [] for key in pipeline_tasks: if key not in in_table: lowercase : List[str] = pipeline_tasks[key]['pt'] if isinstance(__a , (list, tuple) ): lowercase : Any = model[0] lowercase : Union[str, Any] = model.__name__ if model not in in_table.values(): missing.append(__a ) if len(__a ) > 0: lowercase : Optional[Any] = ', '.join(__a ) raise ValueError( "The following pipeline tags are not present in the `PIPELINE_TAGS_AND_AUTO_MODELS` constant inside " f"""`utils/update_metadata.py`: {msg}. Please add them!""" ) if __name__ == "__main__": _A : Any = argparse.ArgumentParser() parser.add_argument("""--token""", type=str, help="""The token to use to push to the transformers-metadata dataset.""") parser.add_argument("""--commit_sha""", type=str, help="""The sha of the commit going with this update.""") parser.add_argument("""--check-only""", action="""store_true""", help="""Activate to just check all pipelines are present.""") _A : List[str] = parser.parse_args() if args.check_only: check_pipeline_tags() else: update_metadata(args.token, args.commit_sha)
202
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = 42 _SCREAMING_SNAKE_CASE = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
224
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case = ''' Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\") >>> pipe.to(\"cuda\") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save(\"cat.png\") ``` ''' def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=8 ): '''simple docstring''' lowerCamelCase : Any = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowerCamelCase : int = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' def __init__( self , __A , __A , __A , ): """simple docstring""" super().__init__() self.register_modules( unet=_a , scheduler=_a , movq=_a , ) lowerCamelCase : Dict = 2 ** (len(self.movq.config.block_out_channels ) - 1) def _snake_case ( self , __A , __A , __A , __A , __A , __A ): """simple docstring""" if latents is None: lowerCamelCase : Optional[int] = randn_tensor(_a , generator=_a , device=_a , dtype=_a ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) lowerCamelCase : Tuple = latents.to(_a ) lowerCamelCase : str = latents * scheduler.init_noise_sigma return latents def _snake_case ( self , __A=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowerCamelCase : int = torch.device(F"""cuda:{gpu_id}""" ) lowerCamelCase : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_a , _a ) def _snake_case ( self , __A=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowerCamelCase : Union[str, Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_a ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowerCamelCase : int = None for cpu_offloaded_model in [self.unet, self.movq]: lowerCamelCase : int = cpu_offload_with_hook(_a , _a , prev_module_hook=_a ) # We'll offload the last model manually. lowerCamelCase : Dict = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _snake_case ( self ): """simple docstring""" if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_a , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_a ) def __call__( self , __A , __A , __A = 512 , __A = 512 , __A = 100 , __A = 4.0 , __A = 1 , __A = None , __A = None , __A = "pil" , __A = True , ): """simple docstring""" lowerCamelCase : List[Any] = self._execution_device lowerCamelCase : List[Any] = guidance_scale > 1.0 if isinstance(_a , _a ): lowerCamelCase : str = torch.cat(_a , dim=0 ) lowerCamelCase : Optional[Any] = image_embeds.shape[0] * num_images_per_prompt if isinstance(_a , _a ): lowerCamelCase : Optional[Any] = torch.cat(_a , dim=0 ) if do_classifier_free_guidance: lowerCamelCase : List[str] = image_embeds.repeat_interleave(_a , dim=0 ) lowerCamelCase : List[Any] = negative_image_embeds.repeat_interleave(_a , dim=0 ) lowerCamelCase : Tuple = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_a ) self.scheduler.set_timesteps(_a , device=_a ) lowerCamelCase : Any = self.scheduler.timesteps lowerCamelCase : List[Any] = self.unet.config.in_channels lowerCamelCase : Any = downscale_height_and_width(_a , _a , self.movq_scale_factor ) # create initial latent lowerCamelCase : List[str] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _a , _a , _a , self.scheduler , ) for i, t in enumerate(self.progress_bar(_a ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase : List[str] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase : Dict = {'image_embeds': image_embeds} lowerCamelCase : Any = self.unet( sample=_a , timestep=_a , encoder_hidden_states=_a , added_cond_kwargs=_a , return_dict=_a , )[0] if do_classifier_free_guidance: lowerCamelCase : str = noise_pred.split(latents.shape[1] , dim=1 ) lowerCamelCase : str = noise_pred.chunk(2 ) lowerCamelCase : Any = variance_pred.chunk(2 ) lowerCamelCase : Union[str, Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowerCamelCase : Optional[Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowerCamelCase : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase : int = self.scheduler.step( _a , _a , _a , generator=_a , )[0] # post-processing lowerCamelCase : int = self.movq.decode(_a , force_not_quantize=_a )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: lowerCamelCase : str = image * 0.5 + 0.5 lowerCamelCase : Union[str, Any] = image.clamp(0 , 1 ) lowerCamelCase : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCamelCase : Tuple = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
283
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
def lowerCamelCase__ ( a__ : int , a__ : int ) -> str: return "\n".join( f'''{number} * {i} = {number * i}''' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
122
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' from __future__ import annotations import time A =list[tuple[int, int]] A =[ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] A =[[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class _a : def __init__( self : Any , lowercase : int , lowercase : int , lowercase : int , lowercase : int , lowercase : Node | None ): '''simple docstring''' UpperCAmelCase = pos_x UpperCAmelCase = pos_y UpperCAmelCase = (pos_y, pos_x) UpperCAmelCase = goal_x UpperCAmelCase = goal_y UpperCAmelCase = parent class _a : def __init__( self : Tuple , lowercase : tuple[int, int] , lowercase : tuple[int, int] ): '''simple docstring''' UpperCAmelCase = Node(start[1] , start[0] , goal[1] , goal[0] , _a ) UpperCAmelCase = Node(goal[1] , goal[0] , goal[1] , goal[0] , _a ) UpperCAmelCase = [self.start] UpperCAmelCase = False def A ( self : Union[str, Any] ): '''simple docstring''' while self.node_queue: UpperCAmelCase = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: UpperCAmelCase = True return self.retrace_path(_a ) UpperCAmelCase = self.get_successors(_a ) for node in successors: self.node_queue.append(_a ) if not self.reached: return [self.start.pos] return None def A ( self : List[Any] , lowercase : Node ): '''simple docstring''' UpperCAmelCase = [] for action in delta: UpperCAmelCase = parent.pos_x + action[1] UpperCAmelCase = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(_a ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(_a , _a , self.target.pos_y , self.target.pos_x , _a ) ) return successors def A ( self : List[str] , lowercase : Node | None ): '''simple docstring''' UpperCAmelCase = node UpperCAmelCase = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase = current_node.parent path.reverse() return path class _a : def __init__( self : int , lowercase : str , lowercase : Optional[int] ): '''simple docstring''' UpperCAmelCase = BreadthFirstSearch(_a , _a ) UpperCAmelCase = BreadthFirstSearch(_a , _a ) UpperCAmelCase = False def A ( self : Tuple ): '''simple docstring''' while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: UpperCAmelCase = self.fwd_bfs.node_queue.pop(0 ) UpperCAmelCase = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: UpperCAmelCase = True return self.retrace_bidirectional_path( _a , _a ) UpperCAmelCase = current_bwd_node UpperCAmelCase = current_fwd_node UpperCAmelCase = { self.fwd_bfs: self.fwd_bfs.get_successors(_a ), self.bwd_bfs: self.bwd_bfs.get_successors(_a ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(_a ) if not self.reached: return [self.fwd_bfs.start.pos] return None def A ( self : Dict , lowercase : Node , lowercase : Node ): '''simple docstring''' UpperCAmelCase = self.fwd_bfs.retrace_path(_a ) UpperCAmelCase = self.bwd_bfs.retrace_path(_a ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() A =(0, 0) A =(len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) A =time.time() A =BreadthFirstSearch(init, goal) A =bfs.search() A =time.time() - start_bfs_time print('Unidirectional BFS computation time : ', bfs_time) A =time.time() A =BidirectionalBreadthFirstSearch(init, goal) A =bd_bfs.search() A =time.time() - start_bd_bfs_time print('Bidirectional BFS computation time : ', bd_bfs_time)
34
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig __UpperCamelCase : Dict = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE: def __init__( self: Optional[int] , UpperCamelCase: Tuple , UpperCamelCase: str ) -> str: snake_case__ = question_encoder snake_case__ = generator snake_case__ = self.question_encoder def lowerCAmelCase_ ( self: int , UpperCamelCase: Optional[int] ) -> str: if os.path.isfile(_a ): raise ValueError(F'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(_a , exist_ok=_a ) snake_case__ = os.path.join(_a , 'question_encoder_tokenizer' ) snake_case__ = os.path.join(_a , 'generator_tokenizer' ) self.question_encoder.save_pretrained(_a ) self.generator.save_pretrained(_a ) @classmethod def lowerCAmelCase_ ( cls: Optional[Any] , UpperCamelCase: List[str] , **UpperCamelCase: Optional[int] ) -> int: from ..auto.tokenization_auto import AutoTokenizer snake_case__ = kwargs.pop('config' , _a ) if config is None: snake_case__ = RagConfig.from_pretrained(_a ) snake_case__ = AutoTokenizer.from_pretrained( _a , config=config.question_encoder , subfolder='question_encoder_tokenizer' ) snake_case__ = AutoTokenizer.from_pretrained( _a , config=config.generator , subfolder='generator_tokenizer' ) return cls(question_encoder=_a , generator=_a ) def __call__( self: int , *UpperCamelCase: List[Any] , **UpperCamelCase: List[str] ) -> Union[str, Any]: return self.current_tokenizer(*_a , **_a ) def lowerCAmelCase_ ( self: int , *UpperCamelCase: Optional[int] , **UpperCamelCase: Union[str, Any] ) -> Union[str, Any]: return self.generator.batch_decode(*_a , **_a ) def lowerCAmelCase_ ( self: Union[str, Any] , *UpperCamelCase: int , **UpperCamelCase: Optional[int] ) -> int: return self.generator.decode(*_a , **_a ) def lowerCAmelCase_ ( self: List[str] ) -> Optional[Any]: snake_case__ = self.question_encoder def lowerCAmelCase_ ( self: Optional[Any] ) -> Optional[int]: snake_case__ = self.generator def lowerCAmelCase_ ( self: Any , UpperCamelCase: List[str] , UpperCamelCase: Optional[List[str]] = None , UpperCamelCase: Optional[int] = None , UpperCamelCase: Optional[int] = None , UpperCamelCase: str = "longest" , UpperCamelCase: str = None , UpperCamelCase: bool = True , **UpperCamelCase: Any , ) -> Tuple: warnings.warn( '`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ' 'regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ' 'context manager to prepare your targets. See the documentation of your specific tokenizer for more ' 'details' , _a , ) if max_length is None: snake_case__ = self.current_tokenizer.model_max_length snake_case__ = self( _a , add_special_tokens=_a , return_tensors=_a , max_length=_a , padding=_a , truncation=_a , **_a , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: snake_case__ = self.current_tokenizer.model_max_length snake_case__ = self( text_target=_a , add_special_tokens=_a , return_tensors=_a , padding=_a , max_length=_a , truncation=_a , **_a , ) snake_case__ = labels['input_ids'] return model_inputs
307
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('9.1.0'): a : Any = { 'linear': PIL.Image.Resampling.BILINEAR, 'bilinear': PIL.Image.Resampling.BILINEAR, 'bicubic': PIL.Image.Resampling.BICUBIC, 'lanczos': PIL.Image.Resampling.LANCZOS, 'nearest': PIL.Image.Resampling.NEAREST, } else: a : Dict = { 'linear': PIL.Image.LINEAR, 'bilinear': PIL.Image.BILINEAR, 'bicubic': PIL.Image.BICUBIC, 'lanczos': PIL.Image.LANCZOS, 'nearest': PIL.Image.NEAREST, } def lowerCAmelCase_ (lowerCAmelCase__: Union[str, Any] ): """simple docstring""" UpperCAmelCase_: Optional[Any] = (images / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase_: Dict = images.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() UpperCAmelCase_: Dict = numpy_to_pil(__a ) return images def lowerCAmelCase_ (lowerCAmelCase__: Dict ): """simple docstring""" if images.ndim == 3: UpperCAmelCase_: Dict = images[None, ...] UpperCAmelCase_: Tuple = (images * 2_5_5).round().astype("""uint8""" ) if images.shape[-1] == 1: # special case for grayscale (single channel) images UpperCAmelCase_: Dict = [Image.fromarray(image.squeeze() , mode="""L""" ) for image in images] else: UpperCAmelCase_: Optional[Any] = [Image.fromarray(__a ) for image in images] return pil_images
147
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed _snake_case : Dict = { 'distilbert': (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), 'roberta': (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), 'bert': (BertConfig, BertForMaskedLM, BertTokenizer), 'gpt2': (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def a_ ( lowerCAmelCase_ : Optional[Any] ): assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Union[str, Any] ): if args.student_type == "roberta": __lowerCAmelCase = False elif args.student_type == "gpt2": __lowerCAmelCase = False def a_ ( lowerCAmelCase_ : Any, lowerCAmelCase_ : Dict ): if args.student_type == "roberta": __lowerCAmelCase = False def a_ ( ): __lowerCAmelCase = argparse.ArgumentParser(description='Training' ) parser.add_argument('--force', action='store_true', help='Overwrite dump_path if it already exists.' ) parser.add_argument( '--dump_path', type=__a, required=__a, help='The output directory (log, checkpoints, parameters, etc.)' ) parser.add_argument( '--data_file', type=__a, required=__a, help='The binarized file (tokenized + tokens_to_ids) and grouped by sequence.', ) parser.add_argument( '--student_type', type=__a, choices=['distilbert', 'roberta', 'gpt2'], required=__a, help='The student type (DistilBERT, RoBERTa).', ) parser.add_argument('--student_config', type=__a, required=__a, help='Path to the student configuration.' ) parser.add_argument( '--student_pretrained_weights', default=__a, type=__a, help='Load student initialization checkpoint.' ) parser.add_argument( '--teacher_type', choices=['bert', 'roberta', 'gpt2'], required=__a, help='Teacher type (BERT, RoBERTa).' ) parser.add_argument('--teacher_name', type=__a, required=__a, help='The teacher model.' ) parser.add_argument('--temperature', default=2.0, type=__a, help='Temperature for the softmax temperature.' ) parser.add_argument( '--alpha_ce', default=0.5, type=__a, help='Linear weight for the distillation loss. Must be >=0.' ) parser.add_argument( '--alpha_mlm', default=0.0, type=__a, help='Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.', ) parser.add_argument('--alpha_clm', default=0.5, type=__a, help='Linear weight for the CLM loss. Must be >=0.' ) parser.add_argument('--alpha_mse', default=0.0, type=__a, help='Linear weight of the MSE loss. Must be >=0.' ) parser.add_argument( '--alpha_cos', default=0.0, type=__a, help='Linear weight of the cosine embedding loss. Must be >=0.' ) parser.add_argument( '--mlm', action='store_true', help='The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.' ) parser.add_argument( '--mlm_mask_prop', default=0.15, type=__a, help='Proportion of tokens for which we need to make a prediction.', ) parser.add_argument('--word_mask', default=0.8, type=__a, help='Proportion of tokens to mask out.' ) parser.add_argument('--word_keep', default=0.1, type=__a, help='Proportion of tokens to keep.' ) parser.add_argument('--word_rand', default=0.1, type=__a, help='Proportion of tokens to randomly replace.' ) parser.add_argument( '--mlm_smoothing', default=0.7, type=__a, help='Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).', ) parser.add_argument('--token_counts', type=__a, help='The token counts in the data_file for MLM.' ) parser.add_argument( '--restrict_ce_to_mask', action='store_true', help='If true, compute the distillation loss only the [MLM] prediction distribution.', ) parser.add_argument( '--freeze_pos_embs', action='store_true', help='Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.', ) parser.add_argument( '--freeze_token_type_embds', action='store_true', help='Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.', ) parser.add_argument('--n_epoch', type=__a, default=3, help='Number of pass on the whole dataset.' ) parser.add_argument('--batch_size', type=__a, default=5, help='Batch size (for each process).' ) parser.add_argument( '--group_by_size', action='store_false', help='If true, group sequences that have similar length into the same batch. Default is true.', ) parser.add_argument( '--gradient_accumulation_steps', type=__a, default=50, help='Gradient accumulation for larger training batches.', ) parser.add_argument('--warmup_prop', default=0.05, type=__a, help='Linear warmup proportion.' ) parser.add_argument('--weight_decay', default=0.0, type=__a, help='Weight decay if we apply some.' ) parser.add_argument('--learning_rate', default=5E-4, type=__a, help='The initial learning rate for Adam.' ) parser.add_argument('--adam_epsilon', default=1E-6, type=__a, help='Epsilon for Adam optimizer.' ) parser.add_argument('--max_grad_norm', default=5.0, type=__a, help='Max gradient norm.' ) parser.add_argument('--initializer_range', default=0.02, type=__a, help='Random initialization range.' ) parser.add_argument( '--fp16', action='store_true', help='Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit', ) parser.add_argument( '--fp16_opt_level', type=__a, default='O1', help=( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].' 'See details at https://nvidia.github.io/apex/amp.html' ), ) parser.add_argument('--n_gpu', type=__a, default=1, help='Number of GPUs in the node.' ) parser.add_argument('--local_rank', type=__a, default=-1, help='Distributed training - Local rank' ) parser.add_argument('--seed', type=__a, default=56, help='Random seed' ) parser.add_argument('--log_interval', type=__a, default=500, help='Tensorboard logging interval.' ) parser.add_argument('--checkpoint_interval', type=__a, default=4000, help='Checkpoint interval.' ) __lowerCAmelCase = parser.parse_args() sanity_checks(__a ) # ARGS # init_gpu_params(__a ) set_seed(__a ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( F"""Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite""" ' itUse `--force` if you want to overwrite it' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(F"""Experiment will be dumped and logged in {args.dump_path}""" ) # SAVE PARAMS # logger.info(F"""Param: {args}""" ) with open(os.path.join(args.dump_path, 'parameters.json' ), 'w' ) as f: json.dump(vars(__a ), __a, indent=4 ) git_log(args.dump_path ) __lowerCAmelCase = MODEL_CLASSES[args.student_type] __lowerCAmelCase = MODEL_CLASSES[args.teacher_type] # TOKENIZER # __lowerCAmelCase = teacher_tokenizer_class.from_pretrained(args.teacher_name ) __lowerCAmelCase = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): __lowerCAmelCase = tokenizer.all_special_tokens.index(__a ) __lowerCAmelCase = tokenizer.all_special_ids[idx] logger.info(F"""Special tokens {special_tok_ids}""" ) __lowerCAmelCase = special_tok_ids __lowerCAmelCase = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(F"""Loading data from {args.data_file}""" ) with open(args.data_file, 'rb' ) as fp: __lowerCAmelCase = pickle.load(__a ) if args.mlm: logger.info(F"""Loading token counts from {args.token_counts} (already pre-computed)""" ) with open(args.token_counts, 'rb' ) as fp: __lowerCAmelCase = pickle.load(__a ) __lowerCAmelCase = np.maximum(__a, 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): __lowerCAmelCase = 0.0 # do not predict special tokens __lowerCAmelCase = torch.from_numpy(__a ) else: __lowerCAmelCase = None __lowerCAmelCase = LmSeqsDataset(params=__a, data=__a ) logger.info('Data loader created.' ) # STUDENT # logger.info(F"""Loading student config from {args.student_config}""" ) __lowerCAmelCase = student_config_class.from_pretrained(args.student_config ) __lowerCAmelCase = True if args.student_pretrained_weights is not None: logger.info(F"""Loading pretrained weights from {args.student_pretrained_weights}""" ) __lowerCAmelCase = student_model_class.from_pretrained(args.student_pretrained_weights, config=__a ) else: __lowerCAmelCase = student_model_class(__a ) if args.n_gpu > 0: student.to(F"""cuda:{args.local_rank}""" ) logger.info('Student loaded.' ) # TEACHER # __lowerCAmelCase = teacher_model_class.from_pretrained(args.teacher_name, output_hidden_states=__a ) if args.n_gpu > 0: teacher.to(F"""cuda:{args.local_rank}""" ) logger.info(F"""Teacher loaded from {args.teacher_name}.""" ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__a, __a ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__a, __a ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() __lowerCAmelCase = Distiller( params=__a, dataset=__a, token_probs=__a, student=__a, teacher=__a ) distiller.train() logger.info('Let\'s go get some drinks.' ) if __name__ == "__main__": main()
284
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu lowerCamelCase = get_tests_dir() + '''/test_data/fsmt/fsmt_val_data.json''' with io.open(filename, '''r''', encoding='''utf-8''') as f: lowerCamelCase = json.load(f) @require_torch class _a ( unittest.TestCase): def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : Dict )-> Tuple: return FSMTTokenizer.from_pretrained(_a ) def UpperCAmelCase__( self : int , _SCREAMING_SNAKE_CASE : List[str] )-> Optional[int]: lowerCAmelCase__ : List[str] = FSMTForConditionalGeneration.from_pretrained(_a ).to(_a ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['''en-ru''', 26.0], ['''ru-en''', 22.0], ['''en-de''', 22.0], ['''de-en''', 29.0], ] ) @slow def UpperCAmelCase__( self : int , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] )-> Optional[int]: lowerCAmelCase__ : int = F'facebook/wmt19-{pair}' lowerCAmelCase__ : Any = self.get_tokenizer(_a ) lowerCAmelCase__ : Any = self.get_model(_a ) lowerCAmelCase__ : Union[str, Any] = bleu_data[pair]['src'] lowerCAmelCase__ : Optional[int] = bleu_data[pair]['tgt'] lowerCAmelCase__ : Any = tokenizer(_a , return_tensors='''pt''' , truncation=_a , padding='''longest''' ).to(_a ) lowerCAmelCase__ : int = model.generate( input_ids=batch.input_ids , num_beams=8 , ) lowerCAmelCase__ : Optional[int] = tokenizer.batch_decode( _a , skip_special_tokens=_a , clean_up_tokenization_spaces=_a ) lowerCAmelCase__ : Optional[Any] = calculate_bleu(_a , _a ) print(_a ) self.assertGreaterEqual(scores['''bleu'''] , _a )
131
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings lowerCAmelCase__ = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class SCREAMING_SNAKE_CASE__ ( lowercase__ ): """simple docstring""" a : bool =field(default=lowercase__ , metadata={"help": "Whether to use SortishSampler or not."} ) a : bool =field( default=lowercase__ , metadata={"help": "Whether to use generate to calculate generative metrics (ROUGE, BLEU)."} ) a : Optional[int] =field( default=lowercase__ , metadata={ "help": ( "The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default " "to the `max_length` value of the model configuration." ) } , ) a : Optional[int] =field( default=lowercase__ , metadata={ "help": ( "The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default " "to the `num_beams` value of the model configuration." ) } , ) a : Optional[Union[str, Path, GenerationConfig]] =field( default=lowercase__ , metadata={ "help": "Model id, file path or url pointing to a GenerationConfig json file, to use during prediction." } , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = super().to_dict() for k, v in d.items(): if isinstance(_a , _a ): lowerCAmelCase : Tuple = v.to_dict() return d
108
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
"""simple docstring""" def lowercase ( _snake_case : int ) ->int: """simple docstring""" __snake_case : int = int(__a ) if n_element < 1: __snake_case : int = ValueError('''a should be a positive number''' ) raise my_error __snake_case : List[Any] = [1] __snake_case : Optional[int] = (0, 0, 0) __snake_case : Tuple = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[str] = input("""Enter the last number (nth term) of the Hamming Number Series: """) print("""Formula of Hamming Number Series => 2^i * 3^j * 5^k""") SCREAMING_SNAKE_CASE : List[str] = hamming(int(n)) print("""-----------------------------------------------------""") print(F'The list with nth numbers is: {hamming_numbers}') print("""-----------------------------------------------------""")
102
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _A : Tuple = get_logger(__name__) class a__ ( enum.Enum ): __lowerCAmelCase = '''all_checks''' __lowerCAmelCase = '''basic_checks''' __lowerCAmelCase = '''no_checks''' class a__ ( lowercase__ ): pass class a__ ( lowercase__ ): pass class a__ ( lowercase__ ): pass class a__ ( lowercase__ ): pass def __magic_name__ ( __snake_case : Optional[dict] , __snake_case : dict , __snake_case : int=None ) -> Dict: if expected_checksums is None: logger.info("Unable to verify checksums." ) return if len(set(__a ) - set(__a ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__a ) - set(__a ) ) ) if len(set(__a ) - set(__a ) ) > 0: raise UnexpectedDownloadedFile(str(set(__a ) - set(__a ) ) ) lowercase : Tuple = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] lowercase : Optional[Any] = ' for ' + verification_name if verification_name is not None else '' if len(__a ) > 0: raise NonMatchingChecksumError( f"""Checksums didn't match{for_verification_name}:\n""" f"""{bad_urls}\n""" "Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error" ) logger.info("All the checksums matched successfully" + for_verification_name ) class a__ ( lowercase__ ): pass class a__ ( lowercase__ ): pass class a__ ( lowercase__ ): pass class a__ ( lowercase__ ): pass def __magic_name__ ( __snake_case : Optional[dict] , __snake_case : dict ) -> Dict: if expected_splits is None: logger.info("Unable to verify splits sizes." ) return if len(set(__a ) - set(__a ) ) > 0: raise ExpectedMoreSplits(str(set(__a ) - set(__a ) ) ) if len(set(__a ) - set(__a ) ) > 0: raise UnexpectedSplits(str(set(__a ) - set(__a ) ) ) lowercase : Optional[Any] = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__a ) > 0: raise NonMatchingSplitsSizesError(str(__a ) ) logger.info("All the splits matched successfully." ) def __magic_name__ ( __snake_case : str , __snake_case : bool = True ) -> List[Any]: if record_checksum: lowercase : int = shaaaa() with open(__a , "rb" ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B"" ): m.update(__a ) lowercase : Tuple = m.hexdigest() else: lowercase : List[Any] = None return {"num_bytes": os.path.getsize(__a ), "checksum": checksum} def __magic_name__ ( __snake_case : Union[str, Any] ) -> Any: if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
202
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
"""simple docstring""" from collections.abc import Generator from math import sin def UpperCamelCase_ ( lowerCAmelCase__ : bytes ) -> Tuple: """simple docstring""" if len(__a ) != 32: raise ValueError('Input must be of length 32' ) lowerCAmelCase_ : List[Any] = b'' for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCamelCase_ ( lowerCAmelCase__ : int ) -> Optional[Any]: """simple docstring""" if i < 0: raise ValueError('Input must be non-negative' ) lowerCAmelCase_ : List[str] = format(__a , '08x' )[-8:] lowerCAmelCase_ : Dict = b'' for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode('utf-8' ) return little_endian_hex def UpperCamelCase_ ( lowerCAmelCase__ : bytes ) -> str: """simple docstring""" lowerCAmelCase_ : List[Any] = b'' for char in message: bit_string += format(__a , '08b' ).encode('utf-8' ) lowerCAmelCase_ : List[str] = format(len(__a ) , '064b' ).encode('utf-8' ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(__a ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def UpperCamelCase_ ( lowerCAmelCase__ : bytes ) -> Tuple: """simple docstring""" if len(__a ) % 512 != 0: raise ValueError('Input must have length that\'s a multiple of 512' ) for pos in range(0 , len(__a ) , 512 ): lowerCAmelCase_ : Optional[Any] = bit_string[pos : pos + 512] lowerCAmelCase_ : Optional[Any] = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def UpperCamelCase_ ( lowerCAmelCase__ : int ) -> Tuple: """simple docstring""" if i < 0: raise ValueError('Input must be non-negative' ) lowerCAmelCase_ : Union[str, Any] = format(__a , '032b' ) lowerCAmelCase_ : Dict = '' for c in i_str: new_str += "1" if c == "0" else "0" return int(__a , 2 ) def UpperCamelCase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : int ) -> str: """simple docstring""" return (a + b) % 2**32 def UpperCamelCase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : int ) -> List[Any]: """simple docstring""" if i < 0: raise ValueError('Input must be non-negative' ) if shift < 0: raise ValueError('Shift must be non-negative' ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCamelCase_ ( lowerCAmelCase__ : bytes ) -> Dict: """simple docstring""" lowerCAmelCase_ : Optional[Any] = preprocess(__a ) lowerCAmelCase_ : List[str] = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states lowerCAmelCase_ : str = 0x6_7_4_5_2_3_0_1 lowerCAmelCase_ : Dict = 0xe_f_c_d_a_b_8_9 lowerCAmelCase_ : Union[str, Any] = 0x9_8_b_a_d_c_f_e lowerCAmelCase_ : List[str] = 0x1_0_3_2_5_4_7_6 lowerCAmelCase_ : str = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(__a ): lowerCAmelCase_ : Union[str, Any] = aa lowerCAmelCase_ : List[str] = ba lowerCAmelCase_ : List[Any] = ca lowerCAmelCase_ : Tuple = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowerCAmelCase_ : Dict = d ^ (b & (c ^ d)) lowerCAmelCase_ : Any = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowerCAmelCase_ : Dict = c ^ (d & (b ^ c)) lowerCAmelCase_ : Union[str, Any] = (5 * i + 1) % 16 elif i <= 47: lowerCAmelCase_ : List[str] = b ^ c ^ d lowerCAmelCase_ : str = (3 * i + 5) % 16 else: lowerCAmelCase_ : Dict = c ^ (b | not_aa(__a )) lowerCAmelCase_ : Dict = (7 * i) % 16 lowerCAmelCase_ : Optional[int] = (f + a + added_consts[i] + block_words[g]) % 2**32 lowerCAmelCase_ : Union[str, Any] = d lowerCAmelCase_ : str = c lowerCAmelCase_ : List[Any] = b lowerCAmelCase_ : Any = sum_aa(__a , left_rotate_aa(__a , shift_amounts[i] ) ) # Add hashed chunk to running total lowerCAmelCase_ : Optional[int] = sum_aa(__a , __a ) lowerCAmelCase_ : Any = sum_aa(__a , __a ) lowerCAmelCase_ : Tuple = sum_aa(__a , __a ) lowerCAmelCase_ : Dict = sum_aa(__a , __a ) lowerCAmelCase_ : str = reformat_hex(__a ) + reformat_hex(__a ) + reformat_hex(__a ) + reformat_hex(__a ) return digest if __name__ == "__main__": import doctest doctest.testmod()
224
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
_snake_case = '''0.18.2''' from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_note_seq_available, is_onnx_available, is_scipy_available, is_torch_available, is_torchsde_available, is_transformers_available, is_transformers_version, is_unidecode_available, logging, ) try: if not is_onnx_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_onnx_objects import * # noqa F403 else: from .pipelines import OnnxRuntimeModel try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_pt_objects import * # noqa F403 else: from .models import ( AutoencoderKL, ControlNetModel, ModelMixin, PriorTransformer, TaFilmDecoder, TransformeraDModel, UNetaDModel, UNetaDConditionModel, UNetaDModel, UNetaDConditionModel, VQModel, ) from .optimization import ( get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, get_scheduler, ) from .pipelines import ( AudioPipelineOutput, ConsistencyModelPipeline, DanceDiffusionPipeline, DDIMPipeline, DDPMPipeline, DiffusionPipeline, DiTPipeline, ImagePipelineOutput, KarrasVePipeline, LDMPipeline, LDMSuperResolutionPipeline, PNDMPipeline, RePaintPipeline, ScoreSdeVePipeline, ) from .schedulers import ( CMStochasticIterativeScheduler, DDIMInverseScheduler, DDIMParallelScheduler, DDIMScheduler, DDPMParallelScheduler, DDPMScheduler, DEISMultistepScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, HeunDiscreteScheduler, IPNDMScheduler, KarrasVeScheduler, KDPMaAncestralDiscreteScheduler, KDPMaDiscreteScheduler, PNDMScheduler, RePaintScheduler, SchedulerMixin, ScoreSdeVeScheduler, UnCLIPScheduler, UniPCMultistepScheduler, VQDiffusionScheduler, ) from .training_utils import EMAModel try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .schedulers import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .schedulers import DPMSolverSDEScheduler try: if not (is_torch_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipelines import ( AltDiffusionImgaImgPipeline, AltDiffusionPipeline, AudioLDMPipeline, CycleDiffusionPipeline, IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ImageTextPipelineOutput, KandinskyImgaImgPipeline, KandinskyInpaintPipeline, KandinskyPipeline, KandinskyPriorPipeline, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaControlnetPipeline, KandinskyVaaImgaImgPipeline, KandinskyVaaInpaintPipeline, KandinskyVaaPipeline, KandinskyVaaPriorEmbaEmbPipeline, KandinskyVaaPriorPipeline, LDMTextToImagePipeline, PaintByExamplePipeline, SemanticStableDiffusionPipeline, ShapEImgaImgPipeline, ShapEPipeline, StableDiffusionAttendAndExcitePipeline, StableDiffusionControlNetImgaImgPipeline, StableDiffusionControlNetInpaintPipeline, StableDiffusionControlNetPipeline, StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionImageVariationPipeline, StableDiffusionImgaImgPipeline, StableDiffusionInpaintPipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionInstructPixaPixPipeline, StableDiffusionLatentUpscalePipeline, StableDiffusionLDMaDPipeline, StableDiffusionModelEditingPipeline, StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPipeline, StableDiffusionPipelineSafe, StableDiffusionPixaPixZeroPipeline, StableDiffusionSAGPipeline, StableDiffusionUpscalePipeline, StableUnCLIPImgaImgPipeline, StableUnCLIPPipeline, TextToVideoSDPipeline, TextToVideoZeroPipeline, UnCLIPImageVariationPipeline, UnCLIPPipeline, UniDiffuserModel, UniDiffuserPipeline, UniDiffuserTextDecoder, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, VideoToVideoSDPipeline, VQDiffusionPipeline, ) try: if not (is_torch_available() and is_transformers_available() and is_invisible_watermark_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_invisible_watermark_objects import * # noqa F403 else: from .pipelines import StableDiffusionXLImgaImgPipeline, StableDiffusionXLPipeline try: if not (is_torch_available() and is_transformers_available() and is_k_diffusion_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipelines import StableDiffusionKDiffusionPipeline try: if not (is_torch_available() and is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_onnx_objects import * # noqa F403 else: from .pipelines import ( OnnxStableDiffusionImgaImgPipeline, OnnxStableDiffusionInpaintPipeline, OnnxStableDiffusionInpaintPipelineLegacy, OnnxStableDiffusionPipeline, OnnxStableDiffusionUpscalePipeline, StableDiffusionOnnxPipeline, ) try: if not (is_torch_available() and is_librosa_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_librosa_objects import * # noqa F403 else: from .pipelines import AudioDiffusionPipeline, Mel try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .pipelines import SpectrogramDiffusionPipeline try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_objects import * # noqa F403 else: from .models.controlnet_flax import FlaxControlNetModel from .models.modeling_flax_utils import FlaxModelMixin from .models.unet_ad_condition_flax import FlaxUNetaDConditionModel from .models.vae_flax import FlaxAutoencoderKL from .pipelines import FlaxDiffusionPipeline from .schedulers import ( FlaxDDIMScheduler, FlaxDDPMScheduler, FlaxDPMSolverMultistepScheduler, FlaxKarrasVeScheduler, FlaxLMSDiscreteScheduler, FlaxPNDMScheduler, FlaxSchedulerMixin, FlaxScoreSdeVeScheduler, ) try: if not (is_flax_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_and_transformers_objects import * # noqa F403 else: from .pipelines import ( FlaxStableDiffusionControlNetPipeline, FlaxStableDiffusionImgaImgPipeline, FlaxStableDiffusionInpaintPipeline, FlaxStableDiffusionPipeline, ) try: if not (is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_note_seq_objects import * # noqa F403 else: from .pipelines import MidiProcessor
283
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _A = logging.get_logger(__name__) _A = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class lowercase_ ( lowercase__ ): A__ : Dict = '''blip_2_vision_model''' def __init__( self , __UpperCamelCase=1_4_0_8 , __UpperCamelCase=6_1_4_4 , __UpperCamelCase=3_9 , __UpperCamelCase=1_6 , __UpperCamelCase=2_2_4 , __UpperCamelCase=1_4 , __UpperCamelCase="gelu" , __UpperCamelCase=0.00_001 , __UpperCamelCase=0.0 , __UpperCamelCase=1e-10 , __UpperCamelCase=True , **__UpperCamelCase , ): """simple docstring""" super().__init__(**_a ) UpperCamelCase_ = hidden_size UpperCamelCase_ = intermediate_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = patch_size UpperCamelCase_ = image_size UpperCamelCase_ = initializer_range UpperCamelCase_ = attention_dropout UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = hidden_act UpperCamelCase_ = qkv_bias @classmethod def lowerCamelCase_ ( cls , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" cls._set_token_in_kwargs(_a ) UpperCamelCase_ = cls.get_config_dict(_a , **_a ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCamelCase_ = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_a , **_a ) class lowercase_ ( lowercase__ ): A__ : Any = '''blip_2_qformer''' def __init__( self , __UpperCamelCase=3_0_5_2_2 , __UpperCamelCase=7_6_8 , __UpperCamelCase=1_2 , __UpperCamelCase=1_2 , __UpperCamelCase=3_0_7_2 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=5_1_2 , __UpperCamelCase=0.02 , __UpperCamelCase=1e-12 , __UpperCamelCase=0 , __UpperCamelCase="absolute" , __UpperCamelCase=2 , __UpperCamelCase=1_4_0_8 , **__UpperCamelCase , ): """simple docstring""" super().__init__(pad_token_id=_a , **_a ) UpperCamelCase_ = vocab_size UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = hidden_act UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = initializer_range UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = position_embedding_type UpperCamelCase_ = cross_attention_frequency UpperCamelCase_ = encoder_hidden_size @classmethod def lowerCamelCase_ ( cls , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" cls._set_token_in_kwargs(_a ) UpperCamelCase_ = cls.get_config_dict(_a , **_a ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCamelCase_ = config_dict['qformer_config'] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_a , **_a ) class lowercase_ ( lowercase__ ): A__ : Optional[Any] = '''blip-2''' A__ : Dict = True def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=3_2 , **__UpperCamelCase ): """simple docstring""" super().__init__(**_a ) if vision_config is None: UpperCamelCase_ = {} logger.info("""vision_config is None. initializing the Blip2VisionConfig with default values.""" ) if qformer_config is None: UpperCamelCase_ = {} logger.info("""qformer_config is None. Initializing the Blip2QFormerConfig with default values.""" ) if text_config is None: UpperCamelCase_ = {} logger.info("""text_config is None. Initializing the text config with default values (`OPTConfig`).""" ) UpperCamelCase_ = BlipaVisionConfig(**_a ) UpperCamelCase_ = BlipaQFormerConfig(**_a ) UpperCamelCase_ = text_config['model_type'] if 'model_type' in text_config else 'opt' UpperCamelCase_ = CONFIG_MAPPING[text_model_type](**_a ) UpperCamelCase_ = self.text_config.tie_word_embeddings UpperCamelCase_ = self.text_config.is_encoder_decoder UpperCamelCase_ = num_query_tokens UpperCamelCase_ = self.vision_config.hidden_size UpperCamelCase_ = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES UpperCamelCase_ = 1.0 UpperCamelCase_ = 0.02 @classmethod def lowerCamelCase_ ( cls , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , **__UpperCamelCase , ): """simple docstring""" return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **_a , ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = copy.deepcopy(self.__dict__ ) UpperCamelCase_ = self.vision_config.to_dict() UpperCamelCase_ = self.qformer_config.to_dict() UpperCamelCase_ = self.text_config.to_dict() UpperCamelCase_ = self.__class__.model_type return output
122
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
'''simple docstring''' import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class _a : def __init__( self : Optional[Any] , lowercase : List[str] , lowercase : str=13 , lowercase : List[Any]=7 , lowercase : List[str]=True , lowercase : Optional[int]=True , lowercase : List[str]=True , lowercase : Dict=True , lowercase : Optional[Any]=99 , lowercase : Any=64 , lowercase : Optional[int]=32 , lowercase : List[str]=5 , lowercase : str=4 , lowercase : Union[str, Any]=37 , lowercase : Union[str, Any]="gelu" , lowercase : Dict=0.1 , lowercase : Optional[Any]=0.1 , lowercase : Dict=512 , lowercase : Union[str, Any]=16 , lowercase : Optional[int]=2 , lowercase : str=0.02 , lowercase : Dict=3 , lowercase : str=4 , lowercase : Tuple=None , ): '''simple docstring''' UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = embedding_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[Any] ): '''simple docstring''' return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) def A ( self : Tuple , lowercase : int , lowercase : str , lowercase : str , lowercase : Union[str, Any] , lowercase : str , lowercase : int , lowercase : Any ): '''simple docstring''' UpperCAmelCase = MegatronBertModel(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a , attention_mask=_a , token_type_ids=_a ) UpperCAmelCase = model(_a , token_type_ids=_a ) UpperCAmelCase = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Union[str, Any] , lowercase : Dict , lowercase : str , lowercase : int , lowercase : str , lowercase : Dict , lowercase : Optional[Any] , lowercase : str ): '''simple docstring''' UpperCAmelCase = MegatronBertForMaskedLM(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Optional[Any] , lowercase : Union[str, Any] , lowercase : Any , lowercase : Optional[Any] , lowercase : int , lowercase : str , lowercase : str , lowercase : List[Any] ): '''simple docstring''' UpperCAmelCase = MegatronBertForCausalLM(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Any , lowercase : List[str] , lowercase : List[str] , lowercase : List[str] , lowercase : List[str] , lowercase : List[str] , lowercase : List[Any] , lowercase : Tuple ): '''simple docstring''' UpperCAmelCase = MegatronBertForNextSentencePrediction(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model( _a , attention_mask=_a , token_type_ids=_a , labels=_a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def A ( self : Tuple , lowercase : List[str] , lowercase : List[Any] , lowercase : Any , lowercase : List[str] , lowercase : str , lowercase : List[str] , lowercase : Any ): '''simple docstring''' UpperCAmelCase = MegatronBertForPreTraining(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model( _a , attention_mask=_a , token_type_ids=_a , labels=_a , next_sentence_label=_a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def A ( self : str , lowercase : str , lowercase : int , lowercase : str , lowercase : Union[str, Any] , lowercase : Union[str, Any] , lowercase : Optional[int] , lowercase : Dict ): '''simple docstring''' UpperCAmelCase = MegatronBertForQuestionAnswering(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model( _a , attention_mask=_a , token_type_ids=_a , start_positions=_a , end_positions=_a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : str , lowercase : int , lowercase : Dict , lowercase : List[str] , lowercase : Union[str, Any] , lowercase : Tuple , lowercase : Union[str, Any] , lowercase : List[str] ): '''simple docstring''' UpperCAmelCase = self.num_labels UpperCAmelCase = MegatronBertForSequenceClassification(_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Tuple , lowercase : Union[str, Any] , lowercase : Dict , lowercase : List[str] , lowercase : Tuple , lowercase : List[str] , lowercase : Any , lowercase : Optional[int] ): '''simple docstring''' UpperCAmelCase = self.num_labels UpperCAmelCase = MegatronBertForTokenClassification(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : Any , lowercase : Optional[int] , lowercase : Optional[Any] , lowercase : Dict , lowercase : int , lowercase : str , lowercase : List[str] , lowercase : Dict ): '''simple docstring''' UpperCAmelCase = self.num_choices UpperCAmelCase = MegatronBertForMultipleChoice(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = model( _a , attention_mask=_a , token_type_ids=_a , labels=_a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict ): '''simple docstring''' UpperCAmelCase = self.prepare_config_and_inputs() ( UpperCAmelCase ) = config_and_inputs UpperCAmelCase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class _a ( lowercase__ , lowercase__ , unittest.TestCase ): __a : str = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) __a : Tuple = ( { '''feature-extraction''': MegatronBertModel, '''fill-mask''': MegatronBertForMaskedLM, '''question-answering''': MegatronBertForQuestionAnswering, '''text-classification''': MegatronBertForSequenceClassification, '''text-generation''': MegatronBertForCausalLM, '''token-classification''': MegatronBertForTokenClassification, '''zero-shot''': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) __a : List[Any] = True # test_resize_embeddings = False __a : List[Any] = False def A ( self : Any , lowercase : List[Any] , lowercase : List[str] , lowercase : List[str]=False ): '''simple docstring''' UpperCAmelCase = super()._prepare_for_class(_a , _a , return_labels=_a ) if return_labels: if model_class in get_values(_a ): UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_a ) UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_a ) return inputs_dict def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = MegatronBertModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=_a , hidden_size=37 ) def A ( self : str ): '''simple docstring''' self.config_tester.run_common_tests() def A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*_a ) def A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*_a ) def A ( self : int ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*_a ) def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*_a ) def A ( self : List[Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*_a ) def A ( self : Any ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*_a ) def A ( self : Any ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*_a ) def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*_a ) def snake_case_ (_a : Tuple ): return torch.tensor( __a , dtype=torch.long , device=__a , ) A =1E-4 @require_torch @require_sentencepiece @require_tokenizers class _a ( unittest.TestCase ): @slow @unittest.skip('''Model is not available.''' ) def A ( self : str ): '''simple docstring''' UpperCAmelCase = 'nvidia/megatron-bert-uncased-345m' if "MYDIR" in os.environ: UpperCAmelCase = os.path.join(os.environ['''MYDIR'''] , _a ) UpperCAmelCase = MegatronBertModel.from_pretrained(_a ) model.to(_a ) model.half() UpperCAmelCase = _long_tensor([[101, 7_110, 1_005, 1_056, 2_023, 11_333, 17_413, 1_029, 102]] ) with torch.no_grad(): UpperCAmelCase = model(_a )[0] UpperCAmelCase = torch.Size((1, 9, 1_024) ) self.assertEqual(output.shape , _a ) UpperCAmelCase = [-0.6040, -0.2517, -0.1025, 0.3420, -0.6758, -0.0017, -0.1089, -0.1990, 0.5728] for ii in range(3 ): for jj in range(3 ): UpperCAmelCase = output[0, ii, jj] UpperCAmelCase = expected[3 * ii + jj] UpperCAmelCase = 'ii={} jj={} a={} b={}'.format(_a , _a , _a , _a ) self.assertTrue(math.isclose(_a , _a , rel_tol=_a , abs_tol=_a ) , msg=_a )
34
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def a_ ( _A , _A , _A=None ) -> List[str]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, f'''{torch_layer} layer.weight does not match''' snake_case__ = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f'''{torch_layer} layer.bias does not match''' snake_case__ = nn.Parameter(__a ) def a_ ( _A , _A , _A ) -> List[str]: """simple docstring""" snake_case__ = np.asarray(weights[0] ) snake_case__ = np.asarray(weights[1] ) snake_case__ = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def a_ ( _A , _A , _A ) -> str: """simple docstring""" snake_case__ = np.asarray(weights[0] ) snake_case__ = np.asarray(weights[1] ) snake_case__ = np.asarray(weights[2] ) snake_case__ = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def a_ ( _A , _A , _A ) -> Optional[Any]: """simple docstring""" snake_case__ = weights[0][0][0] snake_case__ = np.asarray(layer_norm_a[0] ) snake_case__ = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output snake_case__ = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs snake_case__ = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: snake_case__ = intermediate_weights[2] # layernorm 2 snake_case__ = np.asarray(intermediate_weights[0][0] ) snake_case__ = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense snake_case__ = np.asarray(intermediate_weights[1][0] ) snake_case__ = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out snake_case__ = np.asarray(intermediate_weights[4][0] ) snake_case__ = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def a_ ( _A , _A , _A ) -> int: """simple docstring""" snake_case__ = torch_model.reformer # word embeds snake_case__ = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): snake_case__ = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): snake_case__ = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f'''{position_embeddings[emb_idx]} emb does not match''' snake_case__ = nn.Parameter(torch.tensor(__a ) ) snake_case__ = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): snake_case__ = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm snake_case__ = np.asarray(weights[7][0] ) snake_case__ = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings snake_case__ = np.asarray(weights[9][0] ) snake_case__ = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def a_ ( _A , _A , _A ) -> List[Any]: """simple docstring""" snake_case__ = ReformerConfig.from_json_file(__a ) print(f'''Building PyTorch model from configuration: {config}''' ) snake_case__ = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: snake_case__ = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __UpperCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __UpperCamelCase : List[Any] = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
307
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": a : Dict = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--original_config_file', type=str, required=True, help='The YAML config file corresponding to the original architecture.', ) parser.add_argument( '--num_in_channels', default=None, type=int, help='The number of input channels. If `None` number of input channels will be automatically inferred.', ) parser.add_argument( '--image_size', default=512, type=int, help=( 'The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2' ' Base. Use 768 for Stable Diffusion v2.' ), ) parser.add_argument( '--extract_ema', action='store_true', help=( 'Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights' ' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield' ' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.' ), ) parser.add_argument( '--upcast_attention', action='store_true', help=( 'Whether the attention computation should always be upcasted. This is necessary when running stable' ' diffusion 2.1.' ), ) parser.add_argument( '--from_safetensors', action='store_true', help='If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.', ) parser.add_argument( '--to_safetensors', action='store_true', help='Whether to store pipeline in safetensors format or not.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--device', type=str, help='Device to use (e.g. cpu, cuda:0, cuda:1, etc.)') def lowerCAmelCase_ (lowerCAmelCase__: int ): """simple docstring""" if string == "True": return True elif string == "False": return False else: raise ValueError(F'could not parse string as bool {string}' ) parser.add_argument( '--use_linear_projection', help='Override for use linear projection', required=False, type=parse_bool ) parser.add_argument('--cross_attention_dim', help='Override for cross attention_dim', required=False, type=int) a : int = parser.parse_args() a : Tuple = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
147
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
def a_ ( lowerCAmelCase_ : list, lowerCAmelCase_ : list, lowerCAmelCase_ : int ): __lowerCAmelCase = len(__a ) __lowerCAmelCase = [[0] * n for i in range(__a )] for i in range(__a ): __lowerCAmelCase = y_points[i] for i in range(2, __a ): for j in range(__a, __a ): __lowerCAmelCase = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
284
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
from __future__ import annotations from statistics import mean def lowerCamelCase_ ( _a , _a , _a ): """simple docstring""" lowerCAmelCase__ : str = [0] * no_of_processes lowerCAmelCase__ : Optional[int] = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(__a ): lowerCAmelCase__ : Optional[Any] = burst_time[i] lowerCAmelCase__ : list[int] = [] lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : int = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: lowerCAmelCase__ : Optional[int] = [] lowerCAmelCase__ : int = -1 for i in range(__a ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(__a ) if len(__a ) > 0: lowerCAmelCase__ : List[str] = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: lowerCAmelCase__ : Dict = i total_time += burst_time[target_process] completed += 1 lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : Optional[int] = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def lowerCamelCase_ ( _a , _a , _a ): """simple docstring""" lowerCAmelCase__ : Dict = [0] * no_of_processes for i in range(__a ): lowerCAmelCase__ : Tuple = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print('''[TEST CASE 01]''') lowerCamelCase = 4 lowerCamelCase = [2, 5, 3, 7] lowerCamelCase = [0, 0, 0, 0] lowerCamelCase = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowerCamelCase = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print('''PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time''') for i, process_id in enumerate(list(range(1, 5))): print( f'''{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t''' f'''{waiting_time[i]}\t\t\t\t{turn_around_time[i]}''' ) print(f'''\nAverage waiting time = {mean(waiting_time):.5f}''') print(f'''Average turnaround time = {mean(turn_around_time):.5f}''')
131
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
"""simple docstring""" from ... import PretrainedConfig lowerCAmelCase__ = { '''sijunhe/nezha-cn-base''': '''https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( lowercase__ ): """simple docstring""" a : str =NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP a : str ='''nezha''' def __init__( self , snake_case__=21_128 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3_072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=64 , snake_case__=2 , snake_case__=0.02 , snake_case__=1e-12 , snake_case__=0.1 , snake_case__=0 , snake_case__=2 , snake_case__=3 , snake_case__=True , **snake_case__ , ): """simple docstring""" super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) lowerCAmelCase : Union[str, Any] = vocab_size lowerCAmelCase : Tuple = hidden_size lowerCAmelCase : Optional[Any] = num_hidden_layers lowerCAmelCase : Any = num_attention_heads lowerCAmelCase : Union[str, Any] = hidden_act lowerCAmelCase : List[Any] = intermediate_size lowerCAmelCase : int = hidden_dropout_prob lowerCAmelCase : List[str] = attention_probs_dropout_prob lowerCAmelCase : str = max_position_embeddings lowerCAmelCase : Tuple = max_relative_position lowerCAmelCase : int = type_vocab_size lowerCAmelCase : Dict = initializer_range lowerCAmelCase : Tuple = layer_norm_eps lowerCAmelCase : Any = classifier_dropout lowerCAmelCase : Dict = use_cache
108
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : int = 10 def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : str = [1, 2, 3, 4] __snake_case : Dict = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(_a , self.block_size , 0 ) , _a ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] __snake_case : Optional[int] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_a , self.block_size , 0 ) , _a ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] __snake_case : str = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_a , self.block_size , 0 ) , _a ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = 'It was the year of Our Lord one thousand seven hundred and\n seventy-five.\n\nSpiritual revelations were conceded to England at that\n favoured period, as at this.' __snake_case : str = process_story(_a ) self.assertEqual(_a , [] ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[int] = '' __snake_case : Union[str, Any] = process_story(_a ) self.assertEqual(_a , [] ) self.assertEqual(_a , [] ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = ( 'It was the year of Our Lord one thousand seven hundred and ' 'seventy-five\n\nSpiritual revelations were conceded to England ' 'at that favoured period, as at this.\n@highlight\n\nIt was the best of times' ) __snake_case : List[str] = process_story(_a ) __snake_case : List[Any] = [ 'It was the year of Our Lord one thousand seven hundred and seventy-five.', 'Spiritual revelations were conceded to England at that favoured period, as at this.', ] self.assertEqual(_a , _a ) __snake_case : List[Any] = ['It was the best of times.'] self.assertEqual(_a , _a ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = torch.tensor([1, 2, 3, 4] ) __snake_case : int = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(_a , 0 ).numpy() , expected.numpy() ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) __snake_case : int = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(_a , 23 ).numpy() , expected.numpy() ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) __snake_case : Union[str, Any] = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(_a , 1 ).numpy() , expected.numpy() ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = 1_01 __snake_case : List[Any] = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 1_01, 5, 6], [1, 1_01, 3, 4, 1_01, 6]] ) __snake_case : Any = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) __snake_case : Optional[int] = compute_token_type_ids(_a , _a ) np.testing.assert_array_equal(_a , _a )
102
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
"""simple docstring""" from __future__ import annotations import math def __magic_name__ ( __snake_case : int ) -> List[str]: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__a ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _A : Optional[Any] = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def __magic_name__ ( __snake_case : int ) -> int: if not isinstance(__a , __a ): raise ValueError("n must be an integer" ) if n <= 0: raise ValueError("n must be >= 0" ) lowercase : Dict = [] for num in range(len(__a ) ): lowercase : List[Any] = 0 while 2 * i * i <= odd_composites[num]: lowercase : Any = odd_composites[num] - 2 * i * i if is_prime(__a ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(__a ) == n: return list_nums return [] def __magic_name__ ( ) -> List[str]: return compute_nums(1 )[0] if __name__ == "__main__": print(F"{solution() = }")
202
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : Tuple = logging.get_logger(__name__) lowercase__ : str = { """snap-research/efficientformer-l1-300""": ( """https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json""" ), } class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = '''efficientformer''' def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE_ : List[int] = [3, 2, 6, 4] , SCREAMING_SNAKE_CASE_ : List[int] = [4_8, 9_6, 2_2_4, 4_4_8] , SCREAMING_SNAKE_CASE_ : List[bool] = [True, True, True, True] , SCREAMING_SNAKE_CASE_ : int = 4_4_8 , SCREAMING_SNAKE_CASE_ : int = 3_2 , SCREAMING_SNAKE_CASE_ : int = 4 , SCREAMING_SNAKE_CASE_ : int = 7 , SCREAMING_SNAKE_CASE_ : int = 5 , SCREAMING_SNAKE_CASE_ : int = 8 , SCREAMING_SNAKE_CASE_ : int = 4 , SCREAMING_SNAKE_CASE_ : float = 0.0 , SCREAMING_SNAKE_CASE_ : int = 1_6 , SCREAMING_SNAKE_CASE_ : int = 3 , SCREAMING_SNAKE_CASE_ : int = 3 , SCREAMING_SNAKE_CASE_ : int = 3 , SCREAMING_SNAKE_CASE_ : int = 2 , SCREAMING_SNAKE_CASE_ : int = 1 , SCREAMING_SNAKE_CASE_ : float = 0.0 , SCREAMING_SNAKE_CASE_ : int = 1 , SCREAMING_SNAKE_CASE_ : bool = True , SCREAMING_SNAKE_CASE_ : bool = True , SCREAMING_SNAKE_CASE_ : float = 1E-5 , SCREAMING_SNAKE_CASE_ : str = "gelu" , SCREAMING_SNAKE_CASE_ : float = 0.02 , SCREAMING_SNAKE_CASE_ : float = 1E-12 , SCREAMING_SNAKE_CASE_ : int = 2_2_4 , SCREAMING_SNAKE_CASE_ : float = 1E-05 , **SCREAMING_SNAKE_CASE_ : Union[str, Any] , ): super().__init__(**_a ) lowerCAmelCase_ : Union[str, Any] = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : Optional[int] = hidden_sizes lowerCAmelCase_ : int = num_hidden_layers lowerCAmelCase_ : Optional[int] = num_attention_heads lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Optional[int] = num_channels lowerCAmelCase_ : List[str] = depths lowerCAmelCase_ : int = mlp_expansion_ratio lowerCAmelCase_ : List[Any] = downsamples lowerCAmelCase_ : Any = dim lowerCAmelCase_ : Tuple = key_dim lowerCAmelCase_ : Union[str, Any] = attention_ratio lowerCAmelCase_ : Union[str, Any] = resolution lowerCAmelCase_ : Union[str, Any] = pool_size lowerCAmelCase_ : Tuple = downsample_patch_size lowerCAmelCase_ : Tuple = downsample_stride lowerCAmelCase_ : str = downsample_pad lowerCAmelCase_ : str = drop_path_rate lowerCAmelCase_ : List[str] = num_metaad_blocks lowerCAmelCase_ : Any = distillation lowerCAmelCase_ : Any = use_layer_scale lowerCAmelCase_ : Optional[Any] = layer_scale_init_value lowerCAmelCase_ : str = image_size lowerCAmelCase_ : List[str] = batch_norm_eps
224
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor _snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' def __init__( self , *__A , **__A ): """simple docstring""" warnings.warn( "The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use BeitImageProcessor instead." , _a , ) super().__init__(*_a , **_a )
283
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''asapp/sew-d-tiny-100k''': '''https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json''', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class lowercase_ ( lowercase__ ): A__ : str = '''sew-d''' def __init__( self , __UpperCamelCase=3_2 , __UpperCamelCase=7_6_8 , __UpperCamelCase=1_2 , __UpperCamelCase=1_2 , __UpperCamelCase=3_0_7_2 , __UpperCamelCase=2 , __UpperCamelCase=5_1_2 , __UpperCamelCase=2_5_6 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=("p2c", "c2p") , __UpperCamelCase="layer_norm" , __UpperCamelCase="gelu_python" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=1e-7 , __UpperCamelCase=1e-5 , __UpperCamelCase="group" , __UpperCamelCase="gelu" , __UpperCamelCase=(6_4, 1_2_8, 1_2_8, 1_2_8, 1_2_8, 2_5_6, 2_5_6, 2_5_6, 2_5_6, 5_1_2, 5_1_2, 5_1_2, 5_1_2) , __UpperCamelCase=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , __UpperCamelCase=(1_0, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , __UpperCamelCase=False , __UpperCamelCase=1_2_8 , __UpperCamelCase=1_6 , __UpperCamelCase=True , __UpperCamelCase=0.05 , __UpperCamelCase=1_0 , __UpperCamelCase=2 , __UpperCamelCase=0.0 , __UpperCamelCase=1_0 , __UpperCamelCase=0 , __UpperCamelCase="mean" , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=2_5_6 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , **__UpperCamelCase , ): """simple docstring""" super().__init__(**_a , pad_token_id=_a , bos_token_id=_a , eos_token_id=_a ) UpperCamelCase_ = hidden_size UpperCamelCase_ = feat_extract_norm UpperCamelCase_ = feat_extract_activation UpperCamelCase_ = list(_a ) UpperCamelCase_ = list(_a ) UpperCamelCase_ = list(_a ) UpperCamelCase_ = conv_bias UpperCamelCase_ = num_conv_pos_embeddings UpperCamelCase_ = num_conv_pos_embedding_groups UpperCamelCase_ = len(self.conv_dim ) UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = intermediate_size UpperCamelCase_ = squeeze_factor UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = position_buckets UpperCamelCase_ = share_att_key UpperCamelCase_ = relative_attention UpperCamelCase_ = norm_rel_ebd UpperCamelCase_ = list(_a ) UpperCamelCase_ = hidden_act UpperCamelCase_ = num_attention_heads UpperCamelCase_ = hidden_dropout UpperCamelCase_ = attention_dropout UpperCamelCase_ = activation_dropout UpperCamelCase_ = feat_proj_dropout UpperCamelCase_ = final_dropout UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = feature_layer_norm_eps UpperCamelCase_ = initializer_range UpperCamelCase_ = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" f'''but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)''' f'''= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCamelCase_ = apply_spec_augment UpperCamelCase_ = mask_time_prob UpperCamelCase_ = mask_time_length UpperCamelCase_ = mask_time_min_masks UpperCamelCase_ = mask_feature_prob UpperCamelCase_ = mask_feature_length UpperCamelCase_ = mask_feature_min_masks # ctc loss UpperCamelCase_ = ctc_loss_reduction UpperCamelCase_ = ctc_zero_infinity # sequence classification UpperCamelCase_ = use_weighted_layer_sum UpperCamelCase_ = classifier_proj_size @property def lowerCamelCase_ ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
122
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' def snake_case_ (_a : int = 6_0_0_8_5_1_4_7_5_1_4_3 ): try: UpperCAmelCase = int(__a ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) UpperCAmelCase = 1 UpperCAmelCase = 2 while i * i <= n: while n % i == 0: UpperCAmelCase = i n //= i i += 1 if n > 1: UpperCAmelCase = n return int(__a ) if __name__ == "__main__": print(f"""{solution() = }""")
34
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class __SCREAMING_SNAKE_CASE( lowercase__ ): _UpperCAmelCase = 42 _UpperCAmelCase = None def a_ ( _A , _A=0.999 , _A="cosine" , ) -> Optional[int]: """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(_A ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_A ): return math.exp(t * -12.0 ) else: raise ValueError(f'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) snake_case__ = [] for i in range(__a ): snake_case__ = i / num_diffusion_timesteps snake_case__ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__a ) / alpha_bar_fn(__a ) , __a ) ) return torch.tensor(__a , dtype=torch.floataa ) class __SCREAMING_SNAKE_CASE( lowercase__ , lowercase__ ): _UpperCAmelCase = 1 @register_to_config def __init__( self: int , UpperCamelCase: int = 10_00 , UpperCamelCase: float = 0.0_001 , UpperCamelCase: float = 0.02 , UpperCamelCase: str = "linear" , UpperCamelCase: Optional[Union[np.ndarray, List[float]]] = None , UpperCamelCase: bool = True , UpperCamelCase: bool = True , UpperCamelCase: int = 0 , UpperCamelCase: str = "epsilon" , UpperCamelCase: float = 1.0 , **UpperCamelCase: Union[str, Any] , ) -> Optional[Any]: if kwargs.get('set_alpha_to_one' , _a ) is not None: snake_case__ = ( 'The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead.' ) deprecate('set_alpha_to_one' , '1.0.0' , _a , standard_warn=_a ) snake_case__ = kwargs['set_alpha_to_one'] if trained_betas is not None: snake_case__ = torch.tensor(_a , dtype=torch.floataa ) elif beta_schedule == "linear": snake_case__ = torch.linspace(_a , _a , _a , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case__ = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , _a , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case__ = betas_for_alpha_bar(_a ) else: raise NotImplementedError(F'''{beta_schedule} does is not implemented for {self.__class__}''' ) snake_case__ = 1.0 - self.betas snake_case__ = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. snake_case__ = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution snake_case__ = 1.0 # setable values snake_case__ = None snake_case__ = torch.from_numpy(np.arange(0 , _a ).copy().astype(np.intaa ) ) def lowerCAmelCase_ ( self: List[Any] , UpperCamelCase: torch.FloatTensor , UpperCamelCase: Optional[int] = None ) -> List[Any]: return sample def lowerCAmelCase_ ( self: List[str] , UpperCamelCase: int , UpperCamelCase: Union[str, torch.device] = None ) -> Tuple: if num_inference_steps > self.config.num_train_timesteps: raise ValueError( F'''`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:''' F''' {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle''' F''' maximal {self.config.num_train_timesteps} timesteps.''' ) snake_case__ = num_inference_steps snake_case__ = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 snake_case__ = (np.arange(0 , _a ) * step_ratio).round().copy().astype(np.intaa ) snake_case__ = torch.from_numpy(_a ).to(_a ) self.timesteps += self.config.steps_offset def lowerCAmelCase_ ( self: List[str] , UpperCamelCase: torch.FloatTensor , UpperCamelCase: int , UpperCamelCase: torch.FloatTensor , UpperCamelCase: float = 0.0 , UpperCamelCase: bool = False , UpperCamelCase: Optional[torch.FloatTensor] = None , UpperCamelCase: bool = True , ) -> List[str]: snake_case__ = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process snake_case__ = self.alphas_cumprod[timestep] snake_case__ = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) snake_case__ = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": snake_case__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 snake_case__ = model_output elif self.config.prediction_type == "sample": snake_case__ = model_output snake_case__ = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": snake_case__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output snake_case__ = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or''' ' `v_prediction`' ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: snake_case__ = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf snake_case__ = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf snake_case__ = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=_a , pred_original_sample=_a ) def __len__( self: List[str] ) -> Tuple: return self.config.num_train_timesteps
307
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home a : Dict = HUGGINGFACE_HUB_CACHE a : Any = 'config.json' a : List[Any] = 'diffusion_pytorch_model.bin' a : Any = 'diffusion_flax_model.msgpack' a : List[Any] = 'model.onnx' a : Optional[Any] = 'diffusion_pytorch_model.safetensors' a : str = 'weights.pb' a : Optional[Any] = 'https://huggingface.co' a : str = default_cache_path a : int = 'diffusers_modules' a : Tuple = os.getenv('HF_MODULES_CACHE', os.path.join(hf_cache_home, 'modules')) a : Optional[Any] = ['fp16', 'non-ema'] a : Union[str, Any] = '.self_attn'
147
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def a_ ( ): __lowerCAmelCase = ArgumentParser( description=( 'PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes' ) ) # Optional arguments for the launch helper parser.add_argument('--num_cores', type=__a, default=1, help='Number of TPU cores to use (1 or 8).' ) # positional parser.add_argument( 'training_script', type=__a, help=( 'The full path to the single TPU training ' 'program/script to be launched in parallel, ' 'followed by all the arguments for the ' 'training script' ), ) # rest from the training program parser.add_argument('training_script_args', nargs=__a ) return parser.parse_args() def a_ ( ): __lowerCAmelCase = parse_args() # Import training_script as a module. __lowerCAmelCase = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) __lowerCAmelCase = script_fpath.stem __lowerCAmelCase = importlib.import_module(__a ) # Patch sys.argv __lowerCAmelCase = [args.training_script] + args.training_script_args + ['--tpu_num_cores', str(args.num_cores )] xmp.spawn(mod._mp_fn, args=(), nprocs=args.num_cores ) if __name__ == "__main__": main()
284
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
import re from filelock import FileLock try: import nltk lowerCamelCase = True except (ImportError, ModuleNotFoundError): lowerCamelCase = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def lowerCamelCase_ ( _a ): """simple docstring""" re.sub('''<n>''' , '''''' , __a ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__a ) )
131
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar('''KEY''') lowerCAmelCase__ = TypeVar('''VAL''') @dataclass(frozen=lowercase__ , slots=lowercase__ ) class SCREAMING_SNAKE_CASE__ ( Generic[KEY, VAL] ): """simple docstring""" a : KEY a : VAL class SCREAMING_SNAKE_CASE__ ( _Item ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__(_a , _a ) def __bool__( self ): """simple docstring""" return False lowerCAmelCase__ = _DeletedItem() class SCREAMING_SNAKE_CASE__ ( MutableMapping[KEY, VAL] ): """simple docstring""" def __init__( self , snake_case__ = 8 , snake_case__ = 0.75 ): """simple docstring""" lowerCAmelCase : Any = initial_block_size lowerCAmelCase : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 lowerCAmelCase : Optional[Any] = capacity_factor lowerCAmelCase : Optional[Any] = 0 def lowercase__ ( self , snake_case__ ): """simple docstring""" return hash(_a ) % len(self._buckets ) def lowercase__ ( self , snake_case__ ): """simple docstring""" return (ind + 1) % len(self._buckets ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : Any = self._buckets[ind] if not stored: lowerCAmelCase : Optional[int] = _Item(_a , _a ) self._len += 1 return True elif stored.key == key: lowerCAmelCase : Any = _Item(_a , _a ) return True else: return False def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Any = len(self._buckets ) * self._capacity_factor return len(self ) >= int(_a ) def lowercase__ ( self ): """simple docstring""" if len(self._buckets ) <= self._initial_block_size: return False lowerCAmelCase : int = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : Optional[int] = self._buckets lowerCAmelCase : Dict = [None] * new_size lowerCAmelCase : int = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def lowercase__ ( self ): """simple docstring""" self._resize(len(self._buckets ) * 2 ) def lowercase__ ( self ): """simple docstring""" self._resize(len(self._buckets ) // 2 ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : Dict = self._get_bucket_index(_a ) for _ in range(len(self._buckets ) ): yield ind lowerCAmelCase : Union[str, Any] = self._get_next_ind(_a ) def lowercase__ ( self , snake_case__ , snake_case__ ): """simple docstring""" for ind in self._iterate_buckets(_a ): if self._try_set(_a , _a , _a ): break def __setitem__( self , snake_case__ , snake_case__ ): """simple docstring""" if self._is_full(): self._size_up() self._add_item(_a , _a ) def __delitem__( self , snake_case__ ): """simple docstring""" for ind in self._iterate_buckets(_a ): lowerCAmelCase : Union[str, Any] = self._buckets[ind] if item is None: raise KeyError(_a ) if item is _deleted: continue if item.key == key: lowerCAmelCase : Any = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self , snake_case__ ): """simple docstring""" for ind in self._iterate_buckets(_a ): lowerCAmelCase : Tuple = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(_a ) def __len__( self ): """simple docstring""" return self._len def __iter__( self ): """simple docstring""" yield from (item.key for item in self._buckets if item) def __repr__( self ): """simple docstring""" lowerCAmelCase : Dict = ' ,'.join( f"""{item.key}: {item.val}""" for item in self._buckets if item ) return f"""HashMap({val_string})"""
108
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE : Optional[Any] = { """configuration_clap""": [ """CLAP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ClapAudioConfig""", """ClapConfig""", """ClapTextConfig""", ], """processing_clap""": ["""ClapProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Tuple = [ """CLAP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ClapModel""", """ClapPreTrainedModel""", """ClapTextModel""", """ClapTextModelWithProjection""", """ClapAudioModel""", """ClapAudioModelWithProjection""", ] SCREAMING_SNAKE_CASE : Any = ["""ClapFeatureExtractor"""] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
102
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class a__ ( lowercase__ ): def __init__( self , _a , _a = None , _a = None , _a = None , _a = False , _a = False , _a = None , _a = None , **_a , ): super().__init__( _a , split=_a , features=_a , cache_dir=_a , keep_in_memory=_a , streaming=_a , num_proc=_a , **_a , ) lowercase : List[Any] = field lowercase : List[Any] = path_or_paths if isinstance(_a , _a ) else {self.split: path_or_paths} lowercase : int = Json( cache_dir=_a , data_files=_a , features=_a , field=_a , **_a , ) def __magic_name__ ( self ): if self.streaming: lowercase : Tuple = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: lowercase : Any = None lowercase : Optional[Any] = None lowercase : Dict = None lowercase : Dict = None self.builder.download_and_prepare( download_config=_a , download_mode=_a , verification_mode=_a , base_path=_a , num_proc=self.num_proc , ) lowercase : List[Any] = self.builder.as_dataset( split=self.split , verification_mode=_a , in_memory=self.keep_in_memory ) return dataset class a__ : def __init__( self , _a , _a , _a = None , _a = None , **_a , ): if num_proc is not None and num_proc <= 0: raise ValueError(f"""num_proc {num_proc} must be an integer > 0.""" ) lowercase : Union[str, Any] = dataset lowercase : Optional[Any] = path_or_buf lowercase : Optional[int] = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE lowercase : Any = num_proc lowercase : Optional[Any] = 'utf-8' lowercase : Any = to_json_kwargs def __magic_name__ ( self ): lowercase : List[str] = self.to_json_kwargs.pop("path_or_buf" , _a ) lowercase : List[str] = self.to_json_kwargs.pop("orient" , "records" ) lowercase : Dict = self.to_json_kwargs.pop("lines" , True if orient == "records" else False ) lowercase : Optional[Any] = self.to_json_kwargs.pop("index" , False if orient in ["split", "table"] else True ) lowercase : Union[str, Any] = self.to_json_kwargs.pop("compression" , _a ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(f"""`datasets` currently does not support {compression} compression""" ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , "wb" , compression=_a ) as buffer: lowercase : str = self._write(file_obj=_a , orient=_a , lines=_a , index=_a , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( f"""The compression parameter is not supported when writing to a buffer, but compression={compression}""" " was passed. Please provide a local path instead." ) lowercase : List[Any] = self._write( file_obj=self.path_or_buf , orient=_a , lines=_a , index=_a , **self.to_json_kwargs ) return written def __magic_name__ ( self , _a ): lowercase : Optional[Any] = args lowercase : Any = query_table( table=self.dataset.data , key=slice(_a , offset + self.batch_size ) , indices=self.dataset._indices , ) lowercase : List[Any] = batch.to_pandas().to_json( path_or_buf=_a , orient=_a , lines=_a , index=_a , **_a ) if not json_str.endswith("\n" ): json_str += "\n" return json_str.encode(self.encoding ) def __magic_name__ ( self , _a , _a , _a , _a , **_a , ): lowercase : Dict = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating json from Arrow format" , ): lowercase : Any = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(_a ) else: lowercase : Optional[int] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , _a , _a )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating json from Arrow format" , ): written += file_obj.write(_a ) return written
202
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: lowercase__ : List[Any] = None lowercase__ : List[Any] = logging.get_logger(__name__) lowercase__ : Dict = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} lowercase__ : List[Any] = { """vocab_file""": { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model""", """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model""" ), }, """tokenizer_file""": { """google/bigbird-roberta-base""": ( """https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json""" ), """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json""" ), }, } lowercase__ : List[str] = { """google/bigbird-roberta-base""": 4_0_9_6, """google/bigbird-roberta-large""": 4_0_9_6, """google/bigbird-base-trivia-itc""": 4_0_9_6, } lowercase__ : int = """▁""" class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE = BigBirdTokenizer _SCREAMING_SNAKE_CASE = ['''input_ids''', '''attention_mask'''] _SCREAMING_SNAKE_CASE = [] def __init__( self : int , SCREAMING_SNAKE_CASE_ : List[Any]=None , SCREAMING_SNAKE_CASE_ : List[Any]=None , SCREAMING_SNAKE_CASE_ : Tuple="<unk>" , SCREAMING_SNAKE_CASE_ : Tuple="<s>" , SCREAMING_SNAKE_CASE_ : Union[str, Any]="</s>" , SCREAMING_SNAKE_CASE_ : Dict="<pad>" , SCREAMING_SNAKE_CASE_ : Dict="[SEP]" , SCREAMING_SNAKE_CASE_ : Union[str, Any]="[MASK]" , SCREAMING_SNAKE_CASE_ : Optional[Any]="[CLS]" , **SCREAMING_SNAKE_CASE_ : Union[str, Any] , ): lowerCAmelCase_ : int = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else bos_token lowerCAmelCase_ : List[str] = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else eos_token lowerCAmelCase_ : Optional[Any] = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else unk_token lowerCAmelCase_ : Union[str, Any] = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else pad_token lowerCAmelCase_ : Any = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else cls_token lowerCAmelCase_ : Tuple = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else sep_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : List[Any] = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else mask_token super().__init__( _a , tokenizer_file=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , **_a , ) lowerCAmelCase_ : Union[str, Any] = vocab_file lowerCAmelCase_ : Union[str, Any] = False if not self.vocab_file else True def SCREAMING_SNAKE_CASE__ ( self : List[Any] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None ): lowerCAmelCase_ : Tuple = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE__ ( self : List[Any] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None ): lowerCAmelCase_ : str = [self.sep_token_id] lowerCAmelCase_ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE__ ( self : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(_a ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return lowerCAmelCase_ : Optional[Any] = os.path.join( _a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) return (out_vocab_file,)
224
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' __A : Union[str, Any] = ['''vqvae'''] def __init__( self , __A , __A , __A , __A , ): """simple docstring""" super().__init__() self.register_modules(unet=_a , scheduler=_a , mel=_a , vqvae=_a ) def _snake_case ( self ): """simple docstring""" return 50 if isinstance(self.scheduler , _a ) else 1000 @torch.no_grad() def __call__( self , __A = 1 , __A = None , __A = None , __A = 0 , __A = 0 , __A = None , __A = None , __A = 0 , __A = 0 , __A = None , __A = 0 , __A = None , __A = None , __A=True , ): """simple docstring""" lowerCamelCase : Union[str, Any] = steps or self.get_default_steps() self.scheduler.set_timesteps(_a ) lowerCamelCase : Optional[int] = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: lowerCamelCase : Tuple = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: lowerCamelCase : Optional[Any] = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=_a , device=self.device , ) lowerCamelCase : Optional[int] = noise lowerCamelCase : str = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(_a , _a ) lowerCamelCase : List[Any] = self.mel.audio_slice_to_image(_a ) lowerCamelCase : Optional[Any] = np.frombuffer(input_image.tobytes() , dtype="uint8" ).reshape( (input_image.height, input_image.width) ) lowerCamelCase : Union[str, Any] = (input_image / 255) * 2 - 1 lowerCamelCase : Optional[int] = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: lowerCamelCase : Dict = self.vqvae.encode(torch.unsqueeze(_a , 0 ) ).latent_dist.sample( generator=_a )[0] lowerCamelCase : Optional[Any] = self.vqvae.config.scaling_factor * input_images if start_step > 0: lowerCamelCase : str = self.scheduler.add_noise(_a , _a , self.scheduler.timesteps[start_step - 1] ) lowerCamelCase : int = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) lowerCamelCase : Optional[Any] = int(mask_start_secs * pixels_per_second ) lowerCamelCase : Any = int(mask_end_secs * pixels_per_second ) lowerCamelCase : int = self.scheduler.add_noise(_a , _a , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , _a ): lowerCamelCase : Union[str, Any] = self.unet(_a , _a , _a )['sample'] else: lowerCamelCase : List[Any] = self.unet(_a , _a )['sample'] if isinstance(self.scheduler , _a ): lowerCamelCase : Optional[Any] = self.scheduler.step( model_output=_a , timestep=_a , sample=_a , eta=_a , generator=_a , )['prev_sample'] else: lowerCamelCase : str = self.scheduler.step( model_output=_a , timestep=_a , sample=_a , generator=_a , )['prev_sample'] if mask is not None: if mask_start > 0: lowerCamelCase : str = mask[:, step, :, :mask_start] if mask_end > 0: lowerCamelCase : Union[str, Any] = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance lowerCamelCase : Tuple = 1 / self.vqvae.config.scaling_factor * images lowerCamelCase : Any = self.vqvae.decode(_a )['sample'] lowerCamelCase : List[Any] = (images / 2 + 0.5).clamp(0 , 1 ) lowerCamelCase : Union[str, Any] = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() lowerCamelCase : Union[str, Any] = (images * 255).round().astype("uint8" ) lowerCamelCase : Optional[int] = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(_a , mode="RGB" ).convert("L" ) for _ in images) ) lowerCamelCase : int = [self.mel.image_to_audio(_a ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(_a )[:, np.newaxis, :] ) , **ImagePipelineOutput(_a ) ) @torch.no_grad() def _snake_case ( self , __A , __A = 50 ): """simple docstring""" assert isinstance(self.scheduler , _a ) self.scheduler.set_timesteps(_a ) lowerCamelCase : int = np.array( [np.frombuffer(image.tobytes() , dtype="uint8" ).reshape((1, image.height, image.width) ) for image in images] ) lowerCamelCase : Union[str, Any] = (sample / 255) * 2 - 1 lowerCamelCase : str = torch.Tensor(_a ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): lowerCamelCase : Dict = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps lowerCamelCase : List[str] = self.scheduler.alphas_cumprod[t] lowerCamelCase : Dict = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) lowerCamelCase : Any = 1 - alpha_prod_t lowerCamelCase : List[Any] = self.unet(_a , _a )['sample'] lowerCamelCase : List[str] = (1 - alpha_prod_t_prev) ** 0.5 * model_output lowerCamelCase : Dict = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) lowerCamelCase : List[Any] = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def _snake_case ( __A , __A , __A ): """simple docstring""" lowerCamelCase : Union[str, Any] = acos(torch.dot(torch.flatten(_a ) , torch.flatten(_a ) ) / torch.norm(_a ) / torch.norm(_a ) ) return sin((1 - alpha) * theta ) * xa / sin(_a ) + sin(alpha * theta ) * xa / sin(_a )
283
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
import argparse from collections import defaultdict import yaml _A = '''docs/source/en/_toctree.yml''' def lowerCamelCase__ ( a__ : Optional[int] ) -> Dict: UpperCamelCase_ = defaultdict(__a ) UpperCamelCase_ = [] UpperCamelCase_ = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({"""local""": doc["""local"""], """title""": doc["""title"""]} ) else: new_doc_list.append(__a ) UpperCamelCase_ = new_doc_list UpperCamelCase_ = [key for key, value in counts.items() if value > 1] UpperCamelCase_ = [] for duplicate_key in duplicates: UpperCamelCase_ = list({doc["""title"""] for doc in doc_list if doc["""local"""] == duplicate_key} ) if len(__a ) > 1: raise ValueError( f'''{duplicate_key} is present several times in the documentation table of content at ''' """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if """local""" not in counts or counts[doc["""local"""]] == 1] ) UpperCamelCase_ = sorted(__a , key=lambda a__ : s["title"].lower() ) # "overview" gets special treatment and is always first if len(__a ) > 1: raise ValueError("""{doc_list} has two \'overview\' docs which is not allowed.""" ) overview_doc.extend(__a ) # Sort return overview_doc def lowerCamelCase__ ( a__ : int=False ) -> str: with open(__a , encoding="""utf-8""" ) as f: UpperCamelCase_ = yaml.safe_load(f.read() ) # Get to the API doc UpperCamelCase_ = 0 while content[api_idx]["title"] != "API": api_idx += 1 UpperCamelCase_ = content[api_idx]['sections'] # Then to the model doc UpperCamelCase_ = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 UpperCamelCase_ = api_doc[scheduler_idx]['sections'] UpperCamelCase_ = clean_doc_toc(__a ) UpperCamelCase_ = False if new_scheduler_doc != scheduler_doc: UpperCamelCase_ = True if overwrite: UpperCamelCase_ = new_scheduler_doc if diff: if overwrite: UpperCamelCase_ = api_doc with open(__a , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__a , allow_unicode=__a ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) def lowerCamelCase__ ( a__ : int=False ) -> Optional[Any]: with open(__a , encoding="""utf-8""" ) as f: UpperCamelCase_ = yaml.safe_load(f.read() ) # Get to the API doc UpperCamelCase_ = 0 while content[api_idx]["title"] != "API": api_idx += 1 UpperCamelCase_ = content[api_idx]['sections'] # Then to the model doc UpperCamelCase_ = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 UpperCamelCase_ = False UpperCamelCase_ = api_doc[pipeline_idx]['sections'] UpperCamelCase_ = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: UpperCamelCase_ = pipeline_doc['section'] UpperCamelCase_ = clean_doc_toc(__a ) if overwrite: UpperCamelCase_ = new_sub_pipeline_doc new_pipeline_docs.append(__a ) # sort overall pipeline doc UpperCamelCase_ = clean_doc_toc(__a ) if new_pipeline_docs != pipeline_docs: UpperCamelCase_ = True if overwrite: UpperCamelCase_ = new_pipeline_docs if diff: if overwrite: UpperCamelCase_ = api_doc with open(__a , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__a , allow_unicode=__a ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _A = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
122
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class _a ( unittest.TestCase ): __a : int = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __a : int = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def A ( self : Any , lowercase : Any , lowercase : List[str] , lowercase : str ): '''simple docstring''' UpperCAmelCase = TextaTextGenerationPipeline(model=_a , tokenizer=_a ) return generator, ["Something to write", "Something else"] def A ( self : int , lowercase : int , lowercase : Optional[Any] ): '''simple docstring''' UpperCAmelCase = generator('''Something there''' ) self.assertEqual(_a , [{'''generated_text''': ANY(_a )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['''generated_text'''].startswith('''Something there''' ) ) UpperCAmelCase = generator(['''This is great !''', '''Something else'''] , num_return_sequences=2 , do_sample=_a ) self.assertEqual( _a , [ [{'''generated_text''': ANY(_a )}, {'''generated_text''': ANY(_a )}], [{'''generated_text''': ANY(_a )}, {'''generated_text''': ANY(_a )}], ] , ) UpperCAmelCase = generator( ['''This is great !''', '''Something else'''] , num_return_sequences=2 , batch_size=2 , do_sample=_a ) self.assertEqual( _a , [ [{'''generated_text''': ANY(_a )}, {'''generated_text''': ANY(_a )}], [{'''generated_text''': ANY(_a )}, {'''generated_text''': ANY(_a )}], ] , ) with self.assertRaises(_a ): generator(4 ) @require_torch def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = pipeline('''text2text-generation''' , model='''patrickvonplaten/t5-tiny-random''' , framework='''pt''' ) # do_sample=False necessary for reproducibility UpperCAmelCase = generator('''Something there''' , do_sample=_a ) self.assertEqual(_a , [{'''generated_text''': ''''''}] ) UpperCAmelCase = 3 UpperCAmelCase = generator( '''Something there''' , num_return_sequences=_a , num_beams=_a , ) UpperCAmelCase = [ {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': ''}, ] self.assertEqual(_a , _a ) UpperCAmelCase = generator('''This is a test''' , do_sample=_a , num_return_sequences=2 , return_tensors=_a ) self.assertEqual( _a , [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ] , ) UpperCAmelCase = generator.model.config.eos_token_id UpperCAmelCase = '<pad>' UpperCAmelCase = generator( ['''This is a test''', '''This is a second test'''] , do_sample=_a , num_return_sequences=2 , batch_size=2 , return_tensors=_a , ) self.assertEqual( _a , [ [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ], [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ], ] , ) @require_tf def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = pipeline('''text2text-generation''' , model='''patrickvonplaten/t5-tiny-random''' , framework='''tf''' ) # do_sample=False necessary for reproducibility UpperCAmelCase = generator('''Something there''' , do_sample=_a ) self.assertEqual(_a , [{'''generated_text''': ''''''}] )
34
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class __SCREAMING_SNAKE_CASE( unittest.TestCase ): def __init__( self: Union[str, Any] , UpperCamelCase: int , UpperCamelCase: List[Any]=7 , UpperCamelCase: Any=3 , UpperCamelCase: Optional[Any]=18 , UpperCamelCase: Any=30 , UpperCamelCase: List[str]=4_00 , UpperCamelCase: Dict=True , UpperCamelCase: int=None , UpperCamelCase: Optional[Any]=True , UpperCamelCase: str=None , UpperCamelCase: Optional[int]=True , UpperCamelCase: Any=[0.48_145_466, 0.4_578_275, 0.40_821_073] , UpperCamelCase: str=[0.26_862_954, 0.26_130_258, 0.27_577_711] , UpperCamelCase: Optional[int]=True , ) -> str: snake_case__ = size if size is not None else {'height': 2_24, 'width': 2_24} snake_case__ = crop_size if crop_size is not None else {'height': 18, 'width': 18} snake_case__ = parent snake_case__ = batch_size snake_case__ = num_channels snake_case__ = image_size snake_case__ = min_resolution snake_case__ = max_resolution snake_case__ = do_resize snake_case__ = size snake_case__ = do_center_crop snake_case__ = crop_size snake_case__ = do_normalize snake_case__ = image_mean snake_case__ = image_std snake_case__ = do_convert_rgb def lowerCAmelCase_ ( self: Optional[int] ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def lowerCAmelCase_ ( self: Optional[int] , UpperCamelCase: Optional[Any]=False , UpperCamelCase: Union[str, Any]=False , UpperCamelCase: int=False ) -> Dict: assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: snake_case__ = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 2_55 , size=(self.num_channels, self.max_resolution, self.max_resolution) , dtype=np.uinta ) ) else: snake_case__ = [] for i in range(self.batch_size ): snake_case__ = np.random.choice(np.arange(self.min_resolution , self.max_resolution ) , 2 ) image_inputs.append(np.random.randint(2_55 , size=(self.num_channels, width, height) , dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension snake_case__ = [Image.fromarray(np.moveaxis(_a , 0 , -1 ) ) for x in image_inputs] if torchify: snake_case__ = [torch.from_numpy(_a ) for x in image_inputs] return image_inputs @require_torch @require_vision class __SCREAMING_SNAKE_CASE( lowercase__ , unittest.TestCase ): _UpperCAmelCase = ChineseCLIPImageProcessor if is_vision_available() else None def lowerCAmelCase_ ( self: Union[str, Any] ) -> Optional[Any]: snake_case__ = ChineseCLIPImageProcessingTester(self , do_center_crop=_a ) @property def lowerCAmelCase_ ( self: Tuple ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase_ ( self: List[Any] ) -> Tuple: snake_case__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , 'do_resize' ) ) self.assertTrue(hasattr(_a , 'size' ) ) self.assertTrue(hasattr(_a , 'do_center_crop' ) ) self.assertTrue(hasattr(_a , 'center_crop' ) ) self.assertTrue(hasattr(_a , 'do_normalize' ) ) self.assertTrue(hasattr(_a , 'image_mean' ) ) self.assertTrue(hasattr(_a , 'image_std' ) ) self.assertTrue(hasattr(_a , 'do_convert_rgb' ) ) def lowerCAmelCase_ ( self: Tuple ) -> List[Any]: snake_case__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 2_24, 'width': 2_24} ) self.assertEqual(image_processor.crop_size , {'height': 18, 'width': 18} ) snake_case__ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) def lowerCAmelCase_ ( self: List[str] ) -> Any: pass def lowerCAmelCase_ ( self: Union[str, Any] ) -> Optional[int]: snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case__ = self.image_processor_tester.prepare_inputs(equal_resolution=_a ) for image in image_inputs: self.assertIsInstance(_a , Image.Image ) # Test not batched input snake_case__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def lowerCAmelCase_ ( self: List[Any] ) -> Tuple: snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case__ = self.image_processor_tester.prepare_inputs(equal_resolution=_a , numpify=_a ) for image in image_inputs: self.assertIsInstance(_a , np.ndarray ) # Test not batched input snake_case__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def lowerCAmelCase_ ( self: Dict ) -> List[Any]: snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case__ = self.image_processor_tester.prepare_inputs(equal_resolution=_a , torchify=_a ) for image in image_inputs: self.assertIsInstance(_a , torch.Tensor ) # Test not batched input snake_case__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) @require_torch @require_vision class __SCREAMING_SNAKE_CASE( lowercase__ , unittest.TestCase ): _UpperCAmelCase = ChineseCLIPImageProcessor if is_vision_available() else None def lowerCAmelCase_ ( self: str ) -> str: snake_case__ = ChineseCLIPImageProcessingTester(self , num_channels=4 , do_center_crop=_a ) snake_case__ = 3 @property def lowerCAmelCase_ ( self: str ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase_ ( self: Union[str, Any] ) -> Optional[int]: snake_case__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , 'do_resize' ) ) self.assertTrue(hasattr(_a , 'size' ) ) self.assertTrue(hasattr(_a , 'do_center_crop' ) ) self.assertTrue(hasattr(_a , 'center_crop' ) ) self.assertTrue(hasattr(_a , 'do_normalize' ) ) self.assertTrue(hasattr(_a , 'image_mean' ) ) self.assertTrue(hasattr(_a , 'image_std' ) ) self.assertTrue(hasattr(_a , 'do_convert_rgb' ) ) def lowerCAmelCase_ ( self: Tuple ) -> List[Any]: pass def lowerCAmelCase_ ( self: List[str] ) -> Tuple: snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case__ = self.image_processor_tester.prepare_inputs(equal_resolution=_a ) for image in image_inputs: self.assertIsInstance(_a , Image.Image ) # Test not batched input snake_case__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
307
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand a : List[Any] = ( '4S 3H 2C 7S 5H', '9D 8H 2C 6S 7H', '2D 6D 9D TH 7D', 'TC 8C 2S JH 6C', 'JH 8S TH AH QH', 'TS KS 5S 9S AC', 'KD 6S 9D TH AD', 'KS 8D 4D 9S 4S', # pair '8C 4S KH JS 4D', # pair 'QH 8H KD JH 8S', # pair 'KC 4H KS 2H 8D', # pair 'KD 4S KC 3H 8S', # pair 'AH 8S AS KC JH', # pair '3H 4C 4H 3S 2H', # 2 pairs '5S 5D 2C KH KH', # 2 pairs '3C KH 5D 5S KH', # 2 pairs 'AS 3C KH AD KH', # 2 pairs '7C 7S 3S 7H 5S', # 3 of a kind '7C 7S KH 2H 7H', # 3 of a kind 'AC KH QH AH AS', # 3 of a kind '2H 4D 3C AS 5S', # straight (low ace) '3C 5C 4C 2C 6H', # straight '6S 8S 7S 5H 9H', # straight 'JS QS 9H TS KH', # straight 'QC KH TS JS AH', # straight (high ace) '8C 9C 5C 3C TC', # flush '3S 8S 9S 5S KS', # flush '4C 5C 9C 8C KC', # flush 'JH 8H AH KH QH', # flush '3D 2H 3H 2C 2D', # full house '2H 2C 3S 3H 3D', # full house 'KH KC 3S 3H 3D', # full house 'JC 6H JS JD JH', # 4 of a kind 'JC 7H JS JD JH', # 4 of a kind 'JC KH JS JD JH', # 4 of a kind '2S AS 4S 5S 3S', # straight flush (low ace) '2D 6D 3D 4D 5D', # straight flush '5C 6C 3C 7C 4C', # straight flush 'JH 9H TH KH QH', # straight flush 'JH AH TH KH QH', # royal flush (high ace straight flush) ) a : int = ( ('2H 3H 4H 5H 6H', 'KS AS TS QS JS', 'Loss'), ('2H 3H 4H 5H 6H', 'AS AD AC AH JD', 'Win'), ('AS AH 2H AD AC', 'JS JD JC JH 3D', 'Win'), ('2S AH 2H AS AC', 'JS JD JC JH AD', 'Loss'), ('2S AH 2H AS AC', '2H 3H 5H 6H 7H', 'Win'), ('AS 3S 4S 8S 2S', '2H 3H 5H 6H 7H', 'Win'), ('2H 3H 5H 6H 7H', '2S 3H 4H 5S 6C', 'Win'), ('2S 3H 4H 5S 6C', '3D 4C 5H 6H 2S', 'Tie'), ('2S 3H 4H 5S 6C', 'AH AC 5H 6H AS', 'Win'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H AS', 'Loss'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H 7S', 'Win'), ('6S AD 7H 4S AS', 'AH AC 5H 6H 7S', 'Loss'), ('2S AH 4H 5S KC', 'AH AC 5H 6H 7S', 'Loss'), ('2S 3H 6H 7S 9C', '7H 3C TH 6H 9S', 'Loss'), ('4S 5H 6H TS AC', '3S 5H 6H TS AC', 'Win'), ('2S AH 4H 5S 6C', 'AD 4C 5H 6H 2C', 'Tie'), ('AS AH 3H AD AC', 'AS AH 2H AD AC', 'Win'), ('AH AC 5H 5C QS', 'AH AC 5H 5C KS', 'Loss'), ('AH AC 5H 5C QS', 'KH KC 5H 5C QS', 'Win'), ('7C 7S KH 2H 7H', '3C 3S AH 2H 3H', 'Win'), ('3C 3S AH 2H 3H', '7C 7S KH 2H 7H', 'Loss'), ('6H 5H 4H 3H 2H', '5H 4H 3H 2H AH', 'Win'), ('5H 4H 3H 2H AH', '5H 4H 3H 2H AH', 'Tie'), ('5H 4H 3H 2H AH', '6H 5H 4H 3H 2H', 'Loss'), ('AH AD KS KC AC', 'AH KD KH AC KC', 'Win'), ('2H 4D 3C AS 5S', '2H 4D 3C 6S 5S', 'Loss'), ('2H 3S 3C 3H 2S', '3S 3C 2S 2H 2D', 'Win'), ('4D 6D 5D 2D JH', '3S 8S 3H TC KH', 'Loss'), ('4S 6C 8S 3S 7S', 'AD KS 2D 7D 7C', 'Loss'), ('6S 4C 7H 8C 3H', '5H JC AH 9D 9C', 'Loss'), ('9D 9H JH TC QH', '3C 2S JS 5C 7H', 'Win'), ('2H TC 8S AD 9S', '4H TS 7H 2C 5C', 'Win'), ('9D 3S 2C 7S 7C', 'JC TD 3C TC 9H', 'Loss'), ) a : Union[str, Any] = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', True), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', False), ('AS 3S 4S 8S 2S', True), ) a : List[str] = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', False), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', True), ) a : Optional[int] = ( ('2H 4D 3C AS 5S', True, [5, 4, 3, 2, 14]), ('2H 5D 3C AS 5S', False, [14, 5, 5, 3, 2]), ('JH QD KC AS TS', False, [14, 13, 12, 11, 10]), ('9D 3S 2C 7S 7C', False, [9, 7, 7, 3, 2]), ) a : Optional[int] = ( ('JH AH TH KH QH', 0), ('JH 9H TH KH QH', 0), ('JC KH JS JD JH', 7), ('KH KC 3S 3H 3D', 6), ('8C 9C 5C 3C TC', 0), ('JS QS 9H TS KH', 0), ('7C 7S KH 2H 7H', 3), ('3C KH 5D 5S KH', 2), ('QH 8H KD JH 8S', 1), ('2D 6D 9D TH 7D', 0), ) a : Union[str, Any] = ( ('JH AH TH KH QH', 23), ('JH 9H TH KH QH', 22), ('JC KH JS JD JH', 21), ('KH KC 3S 3H 3D', 20), ('8C 9C 5C 3C TC', 19), ('JS QS 9H TS KH', 18), ('7C 7S KH 2H 7H', 17), ('3C KH 5D 5S KH', 16), ('QH 8H KD JH 8S', 15), ('2D 6D 9D TH 7D', 14), ) def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: Optional[Any] = randrange(len(__a ) ), randrange(len(__a ) ) UpperCAmelCase_: Any = ['Loss', 'Tie', 'Win'][(play >= oppo) + (play > oppo)] UpperCAmelCase_: List[Any] = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def lowerCAmelCase_ (lowerCAmelCase__: int = 1_0_0 ): """simple docstring""" return (generate_random_hand() for _ in range(__a )) @pytest.mark.parametrize("""hand, expected""" , __a ) def lowerCAmelCase_ (lowerCAmelCase__: List[Any] , lowerCAmelCase__: Any ): """simple docstring""" assert PokerHand(__a )._is_flush() == expected @pytest.mark.parametrize("""hand, expected""" , __a ) def lowerCAmelCase_ (lowerCAmelCase__: Tuple , lowerCAmelCase__: str ): """simple docstring""" assert PokerHand(__a )._is_straight() == expected @pytest.mark.parametrize("""hand, expected, card_values""" , __a ) def lowerCAmelCase_ (lowerCAmelCase__: Any , lowerCAmelCase__: List[str] , lowerCAmelCase__: Any ): """simple docstring""" UpperCAmelCase_: str = PokerHand(__a ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("""hand, expected""" , __a ) def lowerCAmelCase_ (lowerCAmelCase__: Optional[int] , lowerCAmelCase__: Optional[Any] ): """simple docstring""" assert PokerHand(__a )._is_same_kind() == expected @pytest.mark.parametrize("""hand, expected""" , __a ) def lowerCAmelCase_ (lowerCAmelCase__: int , lowerCAmelCase__: Dict ): """simple docstring""" assert PokerHand(__a )._hand_type == expected @pytest.mark.parametrize("""hand, other, expected""" , __a ) def lowerCAmelCase_ (lowerCAmelCase__: Tuple , lowerCAmelCase__: Optional[Any] , lowerCAmelCase__: Union[str, Any] ): """simple docstring""" assert PokerHand(__a ).compare_with(PokerHand(__a ) ) == expected @pytest.mark.parametrize("""hand, other, expected""" , generate_random_hands() ) def lowerCAmelCase_ (lowerCAmelCase__: Optional[Any] , lowerCAmelCase__: Any , lowerCAmelCase__: Optional[Any] ): """simple docstring""" assert PokerHand(__a ).compare_with(PokerHand(__a ) ) == expected def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: str = [PokerHand(__a ) for hand in SORTED_HANDS] UpperCAmelCase_: List[Any] = poker_hands.copy() shuffle(__a ) UpperCAmelCase_: List[Any] = chain(sorted(__a ) ) for index, hand in enumerate(__a ): assert hand == poker_hands[index] def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: int = [PokerHand("""2D AC 3H 4H 5S""" ), PokerHand("""2S 3H 4H 5S 6C""" )] pokerhands.sort(reverse=__a ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: List[str] = PokerHand("""2C 4S AS 3D 5C""" ) UpperCAmelCase_: List[str] = True UpperCAmelCase_: Optional[int] = [5, 4, 3, 2, 1_4] for _ in range(1_0 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: Union[str, Any] = 0 UpperCAmelCase_: int = os.path.abspath(os.path.dirname(__a ) ) UpperCAmelCase_: Union[str, Any] = os.path.join(__a , """poker_hands.txt""" ) with open(__a ) as file_hand: for line in file_hand: UpperCAmelCase_: Tuple = line[:1_4].strip() UpperCAmelCase_: Optional[int] = line[1_5:].strip() UpperCAmelCase_: Any = PokerHand(__a ), PokerHand(__a ) UpperCAmelCase_: List[str] = player.compare_with(__a ) if output == "Win": answer += 1 assert answer == 3_7_6
147
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _snake_case : Dict = HfApi() _snake_case : List[Any] = {} # fmt: off _snake_case : Any = torch.tensor([ -0.75_15, -1.68_83, 0.24_20, 0.03_00, 0.63_47, 1.34_33, -1.17_43, -3.74_67, 1.23_42, -2.24_85, 0.46_36, 0.80_76, -0.79_91, 0.39_69, 0.84_98, 0.91_89, -1.88_87, -3.35_22, 0.76_39, 0.20_40, 0.62_71, -2.71_48, -1.63_16, 3.08_39, 0.31_86, 0.27_21, -0.97_59, -1.24_61, 2.62_57, 1.35_57 ]) _snake_case : Optional[Any] = torch.tensor([ -2.36_39, -2.53_44, 0.00_54, -0.66_74, 1.59_90, 1.01_58, 0.31_24, -2.14_36, 1.87_95, -2.54_29, -0.15_66, -0.39_73, 1.24_90, 2.64_47, 1.22_83, -0.52_08, -2.81_54, -3.51_19, 2.38_38, 1.20_33, 1.72_01, -2.12_56, -1.45_76, 2.79_48, 2.42_04, -0.97_52, -1.25_46, 0.80_27, 3.27_58, 3.13_65 ]) _snake_case : int = torch.tensor([ -0.65_31, -0.68_91, -0.31_72, -0.53_75, -0.91_40, -0.53_67, -0.11_75, -0.78_69, -0.38_08, -0.45_13, -0.20_98, -0.00_83, 0.31_83, 0.51_40, 0.22_47, -0.13_04, -0.13_02, -0.28_02, -0.20_84, -0.20_25, -0.49_67, -0.48_73, -0.08_61, 0.69_25, 0.02_50, 0.12_90, -0.15_43, 0.63_16, 1.04_60, 1.49_43 ]) _snake_case : Optional[Any] = torch.tensor([ 0.09_11, 0.11_07, 0.01_82, 0.04_35, -0.08_05, -0.06_08, 0.03_81, 0.21_72, -0.02_80, 0.13_27, -0.02_99, -0.02_55, -0.00_50, -0.11_70, -0.10_46, 0.03_09, 0.13_67, 0.17_28, -0.05_33, -0.07_48, -0.05_34, 0.16_24, 0.03_84, -0.18_05, -0.07_07, 0.06_42, 0.02_20, -0.01_34, -0.13_33, -0.15_05 ]) _snake_case : Dict = torch.tensor([ 0.13_21, 0.13_37, 0.04_40, 0.06_22, -0.05_91, -0.03_70, 0.05_03, 0.21_33, -0.01_77, 0.14_15, -0.01_16, -0.01_12, 0.00_44, -0.09_80, -0.07_89, 0.03_95, 0.15_02, 0.17_85, -0.04_88, -0.05_14, -0.04_04, 0.15_39, 0.04_54, -0.15_59, -0.06_65, 0.06_59, 0.03_83, -0.00_05, -0.12_66, -0.13_86 ]) _snake_case : Dict = torch.tensor([ 0.11_54, 0.12_18, 0.03_07, 0.05_26, -0.07_11, -0.05_41, 0.03_66, 0.20_78, -0.02_67, 0.13_17, -0.02_26, -0.01_93, -0.00_14, -0.10_55, -0.09_02, 0.03_30, 0.13_91, 0.17_09, -0.05_62, -0.06_93, -0.05_60, 0.14_82, 0.03_81, -0.16_83, -0.06_81, 0.06_61, 0.03_31, -0.00_46, -0.12_68, -0.14_31 ]) _snake_case : Optional[Any] = torch.tensor([ 0.11_92, 0.12_40, 0.04_14, 0.06_06, -0.05_57, -0.04_12, 0.04_30, 0.20_42, -0.02_00, 0.13_85, -0.01_15, -0.01_32, 0.00_17, -0.09_65, -0.08_02, 0.03_98, 0.14_33, 0.17_47, -0.04_58, -0.05_33, -0.04_07, 0.15_45, 0.04_19, -0.15_74, -0.06_45, 0.06_26, 0.03_41, -0.00_10, -0.11_99, -0.13_90 ]) _snake_case : str = torch.tensor([ 0.10_75, 0.10_74, 0.02_05, 0.04_31, -0.07_74, -0.06_07, 0.02_98, 0.20_42, -0.03_20, 0.12_67, -0.02_81, -0.02_50, -0.00_64, -0.10_91, -0.09_46, 0.02_90, 0.13_28, 0.16_50, -0.05_80, -0.07_38, -0.05_86, 0.14_40, 0.03_37, -0.17_46, -0.07_12, 0.06_05, 0.02_50, -0.00_99, -0.13_16, -0.14_73 ]) _snake_case : List[Any] = torch.tensor([ -1.45_72, -2.04_81, -0.04_14, -0.60_05, 1.41_36, 0.58_48, 0.40_28, -2.73_30, 1.22_12, -2.12_28, 0.21_55, 0.40_39, 0.76_62, 2.05_35, 0.74_77, -0.32_43, -2.17_58, -2.76_48, 1.69_47, 0.70_26, 1.23_38, -1.60_78, -0.86_82, 2.28_10, 1.85_74, -0.57_18, -0.55_86, -0.01_86, 2.34_15, 2.12_51]) _snake_case : int = torch.tensor([ -1.36_90, -1.97_20, -0.40_90, -0.69_66, 1.46_60, 0.99_38, -0.13_85, -2.73_24, 0.77_36, -1.89_17, 0.29_23, 0.42_93, 0.16_93, 1.41_12, 1.18_87, -0.31_81, -2.21_60, -2.63_81, 1.31_70, 0.81_63, 0.92_40, -1.65_44, -0.60_99, 2.52_59, 1.64_30, -0.90_90, -0.93_92, -0.01_26, 2.42_68, 2.32_66 ]) _snake_case : Tuple = torch.tensor([ -1.35_25, -1.96_28, -0.39_56, -0.68_60, 1.46_64, 1.00_14, -0.12_59, -2.72_12, 0.77_72, -1.88_11, 0.29_96, 0.43_88, 0.17_04, 1.40_29, 1.17_01, -0.30_27, -2.20_53, -2.62_87, 1.33_50, 0.81_31, 0.92_74, -1.62_92, -0.60_98, 2.51_31, 1.65_05, -0.89_58, -0.92_98, -0.01_51, 2.42_57, 2.33_55 ]) _snake_case : Optional[int] = torch.tensor([ -2.05_85, -2.78_97, -0.28_50, -0.89_40, 1.90_52, 0.57_02, 0.63_45, -3.89_59, 1.59_32, -3.23_19, 0.19_74, 0.02_87, 1.75_66, 2.65_43, 0.83_87, -0.53_51, -3.27_36, -4.33_75, 2.90_29, 1.63_90, 1.46_40, -2.17_01, -1.90_13, 2.93_41, 3.49_81, -0.62_55, -1.16_44, -0.15_91, 3.70_97, 3.20_66 ]) _snake_case : Union[str, Any] = torch.tensor([ -2.31_39, -2.55_94, -0.01_97, -0.67_85, 1.70_01, 1.16_06, 0.30_75, -2.17_40, 1.80_71, -2.56_30, -0.09_26, -0.38_11, 1.21_16, 2.62_46, 1.27_31, -0.53_98, -2.81_53, -3.61_40, 2.38_93, 1.32_62, 1.62_58, -2.18_56, -1.32_67, 2.83_95, 2.37_79, -1.06_23, -1.24_68, 0.89_59, 3.33_67, 3.22_43 ]) _snake_case : Tuple = torch.tensor([ -2.06_28, -2.76_67, -0.20_89, -0.82_63, 2.05_39, 0.59_92, 0.64_95, -3.83_36, 1.60_25, -3.28_17, 0.17_21, -0.06_33, 1.75_16, 2.70_39, 0.81_00, -0.59_08, -3.21_13, -4.43_43, 2.92_57, 1.36_32, 1.55_62, -2.14_89, -1.98_94, 3.05_60, 3.33_96, -0.73_28, -1.04_17, 0.03_83, 3.70_93, 3.23_43 ]) _snake_case : Tuple = torch.tensor([ -1.45_74, -2.05_69, -0.04_73, -0.61_17, 1.40_18, 0.57_69, 0.41_29, -2.73_44, 1.22_41, -2.13_97, 0.20_00, 0.39_37, 0.76_16, 2.04_53, 0.73_24, -0.33_91, -2.17_46, -2.77_44, 1.69_63, 0.69_21, 1.21_87, -1.61_72, -0.88_77, 2.24_39, 1.84_71, -0.58_39, -0.56_05, -0.04_64, 2.32_50, 2.12_19 ]) # fmt: on _snake_case : Union[str, Any] = api.list_models(filter='diffusers') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": _snake_case : str = '/home/patrick/google_checkpoints/' + mod.modelId.split('/')[-1] print(F"""Started running {mod.modelId}!!!""") if mod.modelId.startswith('CompVis'): _snake_case : Tuple = UNetaDModel.from_pretrained(local_checkpoint, subfolder='unet') else: _snake_case : Union[str, Any] = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) _snake_case : Union[str, Any] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) _snake_case : Dict = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): _snake_case : Dict = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['_'.join('_'.join(mod.modelId.split('/')).split('-'))], atol=1e-3 ) print(F"""{mod.modelId} has passed successfully!!!""")
284
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def lowerCamelCase_ ( _a = 8 ): """simple docstring""" lowerCAmelCase__ : Optional[int] = ascii_letters + digits + punctuation return "".join(secrets.choice(__a ) for _ in range(__a ) ) def lowerCamelCase_ ( _a , _a ): """simple docstring""" i -= len(__a ) lowerCAmelCase__ : Tuple = i // 3 lowerCAmelCase__ : int = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) lowerCAmelCase__ : Dict = ( chars_incl + random(__a , quotient + remainder ) + random(__a , __a ) + random(__a , __a ) ) lowerCAmelCase__ : Union[str, Any] = list(__a ) shuffle(__a ) return "".join(__a ) # random is a generalised function for letters, characters and numbers def lowerCamelCase_ ( _a , _a ): """simple docstring""" return "".join(secrets.choice(__a ) for _ in range(__a ) ) def lowerCamelCase_ ( _a , _a ): """simple docstring""" pass # Put your code here... def lowerCamelCase_ ( _a , _a ): """simple docstring""" pass # Put your code here... def lowerCamelCase_ ( _a , _a ): """simple docstring""" pass # Put your code here... def lowerCamelCase_ ( _a , _a = 8 ): """simple docstring""" if len(__a ) < min_length: # Your Password must be at least 8 characters long return False lowerCAmelCase__ : Any = any(char in ascii_uppercase for char in password ) lowerCAmelCase__ : Tuple = any(char in ascii_lowercase for char in password ) lowerCAmelCase__ : List[str] = any(char in digits for char in password ) lowerCAmelCase__ : List[str] = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def lowerCamelCase_ ( ): """simple docstring""" lowerCAmelCase__ : Any = int(input('''Please indicate the max length of your password: ''' ).strip() ) lowerCAmelCase__ : Any = input( '''Please indicate the characters that must be in your password: ''' ).strip() print('''Password generated:''' , password_generator(__a ) ) print( '''Alternative Password generated:''' , alternative_password_generator(__a , __a ) , ) print('''[If you are thinking of using this passsword, You better save it.]''' ) if __name__ == "__main__": main()
131
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
"""simple docstring""" import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right lowerCAmelCase__ = 256_047 lowerCAmelCase__ = 256_145 @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( lowercase__ , unittest.TestCase ): """simple docstring""" a : Dict =NllbTokenizer a : Tuple =NllbTokenizerFast a : str =True a : Tuple =True a : Optional[Any] ={} def lowercase__ ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase : int = NllbTokenizer(_a , keep_accents=_a ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[str] = NllbTokenizer(_a , keep_accents=_a ) lowerCAmelCase : List[Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(_a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCAmelCase : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) lowerCAmelCase : List[str] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual( _a , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) lowerCAmelCase : List[str] = tokenizer.convert_ids_to_tokens(_a ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[Any] = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-random-nllb', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(_a , **_a ) lowerCAmelCase : List[Any] = self.tokenizer_class.from_pretrained(_a , **_a ) lowerCAmelCase : Optional[int] = tempfile.mkdtemp() lowerCAmelCase : Optional[int] = tokenizer_r.save_pretrained(_a ) lowerCAmelCase : Tuple = tokenizer_p.save_pretrained(_a ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) lowerCAmelCase : Dict = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(_a , _a ) # Checks everything loads correctly in the same way lowerCAmelCase : Tuple = tokenizer_r.from_pretrained(_a ) lowerCAmelCase : int = tokenizer_p.from_pretrained(_a ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_a , _a ) ) shutil.rmtree(_a ) # Save tokenizer rust, legacy_format=True lowerCAmelCase : Tuple = tempfile.mkdtemp() lowerCAmelCase : List[str] = tokenizer_r.save_pretrained(_a , legacy_format=_a ) lowerCAmelCase : str = tokenizer_p.save_pretrained(_a ) # Checks it save with the same files self.assertSequenceEqual(_a , _a ) # Checks everything loads correctly in the same way lowerCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(_a ) lowerCAmelCase : int = tokenizer_p.from_pretrained(_a ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_a , _a ) ) shutil.rmtree(_a ) # Save tokenizer rust, legacy_format=False lowerCAmelCase : int = tempfile.mkdtemp() lowerCAmelCase : Tuple = tokenizer_r.save_pretrained(_a , legacy_format=_a ) lowerCAmelCase : Tuple = tokenizer_p.save_pretrained(_a ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase : List[str] = tokenizer_r.from_pretrained(_a ) lowerCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(_a ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_a , _a ) ) shutil.rmtree(_a ) @require_torch def lowercase__ ( self ): """simple docstring""" if not self.test_seqaseq: return lowerCAmelCase : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Longer text that will definitely require truncation. lowerCAmelCase : Tuple = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for' ' Syria is that \'there is no military solution\' to the nearly five-year conflict and more weapons' ' will only worsen the violence and misery for millions of people.', ] lowerCAmelCase : int = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al' ' Rusiei pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi' ' că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] try: lowerCAmelCase : Dict = tokenizer.prepare_seqaseq_batch( src_texts=_a , tgt_texts=_a , max_length=3 , max_target_length=10 , return_tensors="pt" , src_lang="eng_Latn" , tgt_lang="ron_Latn" , ) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 10 ) # max_target_length will default to max_length if not specified lowerCAmelCase : Tuple = tokenizer.prepare_seqaseq_batch( _a , tgt_texts=_a , max_length=3 , return_tensors="pt" ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 3 ) lowerCAmelCase : Union[str, Any] = tokenizer.prepare_seqaseq_batch( src_texts=_a , max_length=3 , max_target_length=10 , return_tensors="pt" ) self.assertEqual(batch_encoder_only.input_ids.shape[1] , 3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] , 3 ) self.assertNotIn("decoder_input_ids" , _a ) @unittest.skip("Unfortunately way too slow to build a BPE with SentencePiece." ) def lowercase__ ( self ): """simple docstring""" pass def lowercase__ ( self ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase : int = [AddedToken("<special>" , lstrip=_a )] lowerCAmelCase : Optional[Any] = self.rust_tokenizer_class.from_pretrained( _a , additional_special_tokens=_a , **_a ) lowerCAmelCase : Any = tokenizer_r.encode("Hey this is a <special> token" ) lowerCAmelCase : Union[str, Any] = tokenizer_r.encode("<special>" , add_special_tokens=_a )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: lowerCAmelCase : Union[str, Any] = self.rust_tokenizer_class.from_pretrained( _a , additional_special_tokens=_a , **_a , ) lowerCAmelCase : str = self.tokenizer_class.from_pretrained( _a , additional_special_tokens=_a , **_a ) lowerCAmelCase : List[str] = tokenizer_p.encode("Hey this is a <special> token" ) lowerCAmelCase : Union[str, Any] = tokenizer_cr.encode("Hey this is a <special> token" ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" a : Tuple ='''facebook/nllb-200-distilled-600M''' a : Optional[Any] =[ ''' UN Chief Says There Is No Military Solution in Syria''', ''' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.''', ] a : Optional[Any] =[ '''Şeful ONU declară că nu există o soluţie militară în Siria''', '''Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei''' ''' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor''' ''' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.''', ] a : List[Any] =[ 25_60_47, 1_62_97, 13_44_08, 81_65, 24_80_66, 1_47_34, 9_50, 11_35, 10_57_21, 35_73, 83, 2_73_52, 1_08, 4_94_86, 2, ] @classmethod def lowercase__ ( cls ): """simple docstring""" lowerCAmelCase : NllbTokenizer = NllbTokenizer.from_pretrained( cls.checkpoint_name , src_lang="eng_Latn" , tgt_lang="ron_Latn" ) lowerCAmelCase : str = 1 return cls def lowercase__ ( self ): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ace_Arab"] , 256_001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ace_Latn"] , 256_002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["fra_Latn"] , 256_057 ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Tuple = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _a ) def lowercase__ ( self ): """simple docstring""" self.assertIn(_a , self.tokenizer.all_special_ids ) # fmt: off lowerCAmelCase : Optional[int] = [RO_CODE, 4_254, 98_068, 112_923, 39_072, 3_909, 713, 102_767, 26, 17_314, 35_642, 14_683, 33_118, 2_022, 66_987, 2, 256_047] # fmt: on lowerCAmelCase : List[str] = self.tokenizer.decode(_a , skip_special_tokens=_a ) lowerCAmelCase : Optional[int] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_a ) self.assertEqual(_a , _a ) self.assertNotIn(self.tokenizer.eos_token , _a ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Union[str, Any] = ['this is gunna be a long sentence ' * 20] assert isinstance(src_text[0] , _a ) lowerCAmelCase : Any = 10 lowerCAmelCase : Optional[Any] = self.tokenizer(_a , max_length=_a , truncation=_a ).input_ids[0] self.assertEqual(ids[-1] , 2 ) self.assertEqual(ids[0] , _a ) self.assertEqual(len(_a ) , _a ) def lowercase__ ( self ): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"] ) , [256_203, 3] ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = tempfile.mkdtemp() lowerCAmelCase : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_a ) lowerCAmelCase : List[Any] = NllbTokenizer.from_pretrained(_a ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _a ) @require_torch def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Any = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_a , truncation=_a , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) lowerCAmelCase : Optional[Any] = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.lang_code_to_id["ron_Latn"] ) self.assertIsInstance(_a , _a ) self.assertEqual((2, 15) , batch.input_ids.shape ) self.assertEqual((2, 15) , batch.attention_mask.shape ) lowerCAmelCase : Any = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _a ) self.assertEqual(_a , batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = self.tokenizer(self.src_text , padding=_a , truncation=_a , max_length=3 , return_tensors="pt" ) lowerCAmelCase : Optional[int] = self.tokenizer( text_target=self.tgt_text , padding=_a , truncation=_a , max_length=10 , return_tensors="pt" ) lowerCAmelCase : Optional[int] = targets['input_ids'] lowerCAmelCase : List[str] = shift_tokens_right( _a , self.tokenizer.pad_token_id , decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] , ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Union[str, Any] = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( nested_simplify(_a ) , { # A, test, EOS, en_XX "input_ids": [[256_047, 70, 7_356, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 256_057, } , ) @require_torch def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : str = True lowerCAmelCase : Tuple = self.tokenizer( "UN Chief says there is no military solution in Syria" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( inputs.input_ids , [16_297, 134_408, 25_653, 6_370, 248, 254, 103_929, 94_995, 108, 49_486, 2, 256_047] ) lowerCAmelCase : Tuple = False lowerCAmelCase : Dict = self.tokenizer( "UN Chief says there is no military solution in Syria" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( inputs.input_ids , [256_047, 16_297, 134_408, 25_653, 6_370, 248, 254, 103_929, 94_995, 108, 49_486, 2] )
108
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" def lowercase ( _snake_case : list[int] ) ->List[Any]: """simple docstring""" if not numbers: return 0 if not isinstance(__a , (list, tuple) ) or not all( isinstance(__a , __a ) for number in numbers ): raise ValueError('''numbers must be an iterable of integers''' ) __snake_case : Any = numbers[0] for i in range(1 , len(__a ) ): # update the maximum and minimum subarray products __snake_case : str = numbers[i] if number < 0: __snake_case : List[Any] = min_till_now, max_till_now __snake_case : Optional[int] = max(__a , max_till_now * number ) __snake_case : Tuple = min(__a , min_till_now * number ) # update the maximum product found till now __snake_case : Union[str, Any] = max(__a , __a ) return max_prod
102
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser _A : Tuple = re.compile(r"""\s+""") def __magic_name__ ( __snake_case : Any ) -> Optional[Any]: return {"hash": hashlib.mda(re.sub(__a , "" , example["content"] ).encode("utf-8" ) ).hexdigest()} def __magic_name__ ( __snake_case : str ) -> Optional[Any]: lowercase : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def __magic_name__ ( __snake_case : str ) -> Union[str, Any]: lowercase : Union[str, Any] = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def __magic_name__ ( __snake_case : Optional[int] , __snake_case : Any ) -> Union[str, Any]: if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def __magic_name__ ( __snake_case : int , __snake_case : Union[str, Any]=5 ) -> Optional[Any]: lowercase : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] lowercase : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def __magic_name__ ( __snake_case : List[str] , __snake_case : Dict=5 , __snake_case : Tuple=0.05 ) -> int: lowercase : Optional[int] = ['unit tests', 'test file', 'configuration file'] lowercase : int = example['content'].splitlines() lowercase : int = 0 lowercase : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test lowercase : int = example['content'].count("\n" ) lowercase : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def __magic_name__ ( __snake_case : Optional[int] ) -> Any: lowercase : List[str] = ['def ', 'class ', 'for ', 'while '] lowercase : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def __magic_name__ ( __snake_case : int , __snake_case : Any=4 ) -> str: lowercase : List[str] = example['content'].splitlines() lowercase : Dict = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def __magic_name__ ( __snake_case : str ) -> Optional[int]: lowercase : Optional[Any] = tokenizer(example["content"] , truncation=__a )['input_ids'] lowercase : Optional[int] = len(example["content"] ) / len(__a ) return {"ratio": ratio} def __magic_name__ ( __snake_case : str ) -> Optional[Any]: lowercase : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def __magic_name__ ( __snake_case : Any , __snake_case : Any , __snake_case : str ) -> int: if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def __magic_name__ ( __snake_case : Union[str, Any] ) -> int: with open(__a , "rb" ) as f_in: with gzip.open(str(__a ) + ".gz" , "wb" , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings _A : Tuple = HfArgumentParser(PreprocessingArguments) _A : List[Any] = parser.parse_args() if args.num_workers is None: _A : List[str] = multiprocessing.cpu_count() _A : int = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset _A : List[str] = time.time() _A : Dict = load_dataset(args.dataset_name, split="""train""") print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing _A : Optional[Any] = time.time() _A : str = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes _A : Any = set(ds.unique("""hash""")) _A : Union[str, Any] = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics _A : Dict = time.time() _A : Dict = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: _A : Optional[int] = time.time() _A , _A : Optional[Any] = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file _A : Optional[int] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) _A : Tuple = output_dir / """data""" data_dir.mkdir(exist_ok=True) _A : str = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): _A : Optional[int] = str(data_dir / F"file-{file_number+1:012}.json") _A : Any = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
202
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class UpperCamelCase__ ( lowercase__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = '''Speech2TextFeatureExtractor''' _SCREAMING_SNAKE_CASE = '''Speech2TextTokenizer''' def __init__( self : Tuple , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Any ): super().__init__(_a , _a ) lowerCAmelCase_ : str = self.feature_extractor lowerCAmelCase_ : List[str] = False def __call__( self : List[str] , *SCREAMING_SNAKE_CASE_ : str , **SCREAMING_SNAKE_CASE_ : List[Any] ): if self._in_target_context_manager: return self.current_processor(*_a , **_a ) if "raw_speech" in kwargs: warnings.warn('Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.' ) lowerCAmelCase_ : Optional[int] = kwargs.pop('raw_speech' ) else: lowerCAmelCase_ : Optional[Any] = kwargs.pop('audio' , _a ) lowerCAmelCase_ : Any = kwargs.pop('sampling_rate' , _a ) lowerCAmelCase_ : str = kwargs.pop('text' , _a ) if len(_a ) > 0: lowerCAmelCase_ : str = args[0] lowerCAmelCase_ : Dict = args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if audio is not None: lowerCAmelCase_ : Optional[Any] = self.feature_extractor(_a , *_a , sampling_rate=_a , **_a ) if text is not None: lowerCAmelCase_ : Any = self.tokenizer(_a , **_a ) if text is None: return inputs elif audio is None: return encodings else: lowerCAmelCase_ : int = encodings['input_ids'] return inputs def SCREAMING_SNAKE_CASE__ ( self : List[str] , *SCREAMING_SNAKE_CASE_ : Optional[Any] , **SCREAMING_SNAKE_CASE_ : Union[str, Any] ): return self.tokenizer.batch_decode(*_a , **_a ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , *SCREAMING_SNAKE_CASE_ : Optional[Any] , **SCREAMING_SNAKE_CASE_ : Optional[Any] ): return self.tokenizer.decode(*_a , **_a ) @contextmanager def SCREAMING_SNAKE_CASE__ ( self : List[str] ): warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your audio inputs, or in a separate call.' ) lowerCAmelCase_ : int = True lowerCAmelCase_ : Dict = self.tokenizer yield lowerCAmelCase_ : Any = self.feature_extractor lowerCAmelCase_ : Any = False
224
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel _snake_case = logging.getLogger(__name__) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if os.path.exists(__a ): if os.path.exists(os.path.join(__a , "config.json" ) ) and os.path.isfile( os.path.join(__a , "config.json" ) ): os.remove(os.path.join(__a , "config.json" ) ) if os.path.exists(os.path.join(__a , "pytorch_model.bin" ) ) and os.path.isfile( os.path.join(__a , "pytorch_model.bin" ) ): os.remove(os.path.join(__a , "pytorch_model.bin" ) ) else: os.makedirs(__a ) model.save_pretrained(__a ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : Optional[int] = 2 if unlogit: lowerCamelCase : List[str] = torch.pow(__a , __a ) lowerCamelCase : Union[str, Any] = p * torch.log(__a ) lowerCamelCase : int = 0 return -plogp.sum(dim=-1 ) def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' logger.info("lv, h >\t" + "\t".join(f"""{x + 1}""" for x in range(len(__a ) ) ) ) for row in range(len(__a ) ): if tensor.dtype != torch.long: logger.info(f"""layer {row + 1}:\t""" + "\t".join(f"""{x:.5f}""" for x in tensor[row].cpu().data ) ) else: logger.info(f"""layer {row + 1}:\t""" + "\t".join(f"""{x:d}""" for x in tensor[row].cpu().data ) ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : Optional[Any] = model.config.num_hidden_layers, model.config.num_attention_heads lowerCamelCase : Optional[int] = torch.zeros(__a , __a ).to(args.device ) lowerCamelCase : Any = torch.zeros(__a , __a ).to(args.device ) if head_mask is None: lowerCamelCase : Union[str, Any] = torch.ones(__a , __a ).to(args.device ) head_mask.requires_grad_(requires_grad=__a ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: lowerCamelCase : List[str] = None lowerCamelCase : List[Any] = 0.0 lowerCamelCase : List[str] = 0.0 for step, inputs in enumerate(tqdm(__a , desc="Iteration" , disable=args.local_rank not in [-1, 0] ) ): lowerCamelCase : Optional[int] = tuple(t.to(args.device ) for t in inputs ) (lowerCamelCase ) : List[str] = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) lowerCamelCase : Any = model(__a , labels=__a , head_mask=__a ) # (loss), lm_logits, presents, (all hidden_states), (attentions) lowerCamelCase : Union[str, Any] = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(__a ): lowerCamelCase : List[Any] = entropy(attn.detach() , __a ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(__a ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: lowerCamelCase : str = 2 lowerCamelCase : Any = torch.pow(torch.pow(__a , __a ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1E-20 if not args.dont_normalize_global_importance: lowerCamelCase : str = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info("Attention entropies" ) print_ad_tensor(__a ) if compute_importance: logger.info("Head importance scores" ) print_ad_tensor(__a ) logger.info("Head ranked by importance scores" ) lowerCamelCase : Optional[Any] = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) lowerCamelCase : Any = torch.arange( head_importance.numel() , device=args.device ) lowerCamelCase : Union[str, Any] = head_ranks.view_as(__a ) print_ad_tensor(__a ) return attn_entropy, head_importance, total_loss def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : str = compute_heads_importance(__a , __a , __a , compute_entropy=__a ) lowerCamelCase : List[Any] = 1 / loss # instead of downsteam score use the LM loss logger.info("Pruning: original score: %f, threshold: %f" , __a , original_score * args.masking_threshold ) lowerCamelCase : Any = torch.ones_like(__a ) lowerCamelCase : Union[str, Any] = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) lowerCamelCase : List[str] = original_score while current_score >= original_score * args.masking_threshold: lowerCamelCase : str = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads lowerCamelCase : Tuple = float("Inf" ) lowerCamelCase : Any = head_importance.view(-1 ).sort()[1] if len(__a ) <= num_to_mask: print("BREAK BY num_to_mask" ) break # mask heads lowerCamelCase : int = current_heads_to_mask[:num_to_mask] logger.info("Heads to mask: %s" , str(current_heads_to_mask.tolist() ) ) lowerCamelCase : Tuple = new_head_mask.view(-1 ) lowerCamelCase : str = 0.0 lowerCamelCase : List[Any] = new_head_mask.view_as(__a ) lowerCamelCase : Any = new_head_mask.clone().detach() print_ad_tensor(__a ) # Compute metric and head importance again lowerCamelCase : Tuple = compute_heads_importance( __a , __a , __a , compute_entropy=__a , head_mask=__a ) lowerCamelCase : List[str] = 1 / loss logger.info( "Masking: current score: %f, remaining heads %d (%.1f percents)" , __a , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info("Final head mask" ) print_ad_tensor(__a ) np.save(os.path.join(args.output_dir , "head_mask.npy" ) , head_mask.detach().cpu().numpy() ) return head_mask def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[Any] = datetime.now() lowerCamelCase : Any = compute_heads_importance( __a , __a , __a , compute_entropy=__a , compute_importance=__a , head_mask=__a ) lowerCamelCase : List[Any] = 1 / loss lowerCamelCase : Union[str, Any] = datetime.now() - before_time lowerCamelCase : List[Any] = sum(p.numel() for p in model.parameters() ) lowerCamelCase : Tuple = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(__a ) ) } for k, v in heads_to_prune.items(): if isinstance(__a , __a ): lowerCamelCase : List[str] = [ v, ] assert sum(len(__a ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(__a ) lowerCamelCase : Optional[Any] = sum(p.numel() for p in model.parameters() ) lowerCamelCase : Union[str, Any] = datetime.now() lowerCamelCase : Any = compute_heads_importance( __a , __a , __a , compute_entropy=__a , compute_importance=__a , head_mask=__a , actually_pruned=__a , ) lowerCamelCase : int = 1 / loss lowerCamelCase : Any = datetime.now() - before_time logger.info( "Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)" , __a , __a , pruned_num_params / original_num_params * 100 , ) logger.info("Pruning: score with masking: %f score with pruning: %f" , __a , __a ) logger.info("Pruning: speed ratio (original timing / new timing): %f percents" , original_time / new_time * 100 ) save_model(__a , args.output_dir ) def lowercase_( ): '''simple docstring''' lowerCamelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--data_dir" , default=__a , type=__a , required=__a , help="The input data dir. Should contain the .tsv files (or other data files) for the task." , ) parser.add_argument( "--model_name_or_path" , default=__a , type=__a , required=__a , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--output_dir" , default=__a , type=__a , required=__a , help="The output directory where the model predictions and checkpoints will be written." , ) # Other parameters parser.add_argument( "--config_name" , default="" , type=__a , help="Pretrained config name or path if not the same as model_name_or_path" , ) parser.add_argument( "--tokenizer_name" , default="" , type=__a , help="Pretrained tokenizer name or path if not the same as model_name_or_path" , ) parser.add_argument( "--cache_dir" , default=__a , type=__a , help="Where do you want to store the pre-trained models downloaded from s3" , ) parser.add_argument( "--data_subset" , type=__a , default=-1 , help="If > 0: limit the data to a subset of data_subset instances." ) parser.add_argument( "--overwrite_output_dir" , action="store_true" , help="Whether to overwrite data in output directory" ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) parser.add_argument( "--dont_normalize_importance_by_layer" , action="store_true" , help="Don\'t normalize importance score by layers" ) parser.add_argument( "--dont_normalize_global_importance" , action="store_true" , help="Don\'t normalize all importance scores between 0 and 1" , ) parser.add_argument( "--try_masking" , action="store_true" , help="Whether to try to mask head until a threshold of accuracy." ) parser.add_argument( "--masking_threshold" , default=0.9 , type=__a , help="masking threshold in term of metrics (stop masking when metric < threshold * original metric value)." , ) parser.add_argument( "--masking_amount" , default=0.1 , type=__a , help="Amount to heads to masking at each masking step." ) parser.add_argument("--metric_name" , default="acc" , type=__a , help="Metric to use for head masking." ) parser.add_argument( "--max_seq_length" , default=128 , type=__a , help=( "The maximum total input sequence length after WordPiece tokenization. \n" "Sequences longer than this will be truncated, sequences shorter padded." ) , ) parser.add_argument("--batch_size" , default=1 , type=__a , help="Batch size." ) parser.add_argument("--seed" , type=__a , default=42 ) parser.add_argument("--local_rank" , type=__a , default=-1 , help="local_rank for distributed training on gpus" ) parser.add_argument("--no_cuda" , action="store_true" , help="Whether not to use CUDA when available" ) parser.add_argument("--server_ip" , type=__a , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=__a , default="" , help="Can be used for distant debugging." ) lowerCamelCase : str = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=__a ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: lowerCamelCase : Optional[Any] = torch.device("cuda" if torch.cuda.is_available() and not args.no_cuda else "cpu" ) lowerCamelCase : List[str] = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) lowerCamelCase : int = torch.device("cuda" , args.local_rank ) lowerCamelCase : Dict = 1 torch.distributed.init_process_group(backend="nccl" ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) lowerCamelCase : List[Any] = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: lowerCamelCase : Any = nn.parallel.DistributedDataParallel( __a , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=__a ) elif args.n_gpu > 1: lowerCamelCase : Union[str, Any] = nn.DataParallel(__a ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=__a ) torch.save(__a , os.path.join(args.output_dir , "run_args.bin" ) ) logger.info("Training/evaluation parameters %s" , __a ) # Prepare dataset lowerCamelCase : Optional[Any] = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) lowerCamelCase : Optional[int] = (torch.from_numpy(__a ),) lowerCamelCase : Optional[int] = TensorDataset(*__a ) lowerCamelCase : List[str] = RandomSampler(__a ) lowerCamelCase : int = DataLoader(__a , sampler=__a , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(__a , __a , __a ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: lowerCamelCase : Optional[int] = mask_heads(__a , __a , __a ) prune_heads(__a , __a , __a , __a ) if __name__ == "__main__": main()
283
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed _A = logging.getLogger(__name__) def lowerCamelCase__ ( a__ : Optional[int]=2 , a__ : Optional[int]=3 , a__ : int=16 , a__ : int = 10 , a__ : int = 2 ) -> Union[str, Any]: def get_dataset(a__ : Any ): UpperCamelCase_ = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(__a , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) UpperCamelCase_ = get_dataset(__a ) UpperCamelCase_ = get_dataset(__a ) UpperCamelCase_ = DataLoader(__a , shuffle=__a , batch_size=__a , num_workers=4 ) UpperCamelCase_ = DataLoader(__a , shuffle=__a , batch_size=__a , num_workers=4 ) return (train_dataloader, valid_dataloader) def lowerCamelCase__ ( a__ : List[str] , a__ : int , a__ : str , a__ : Union[str, Any] , a__ : Any , a__ : Optional[Any]=None ) -> str: UpperCamelCase_ = [] for epoch in range(__a ): # Train quickly model.train() for batch in dataloader: UpperCamelCase_ = batch UpperCamelCase_ = model(__a ) UpperCamelCase_ = torch.nn.functional.mse_loss(__a , __a ) accelerator.backward(__a ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class lowercase_ ( nn.Module ): def __init__( self ): """simple docstring""" super().__init__() UpperCamelCase_ = nn.Parameter(torch.randn(1 ) ) UpperCamelCase_ = nn.Parameter(torch.randn(1 ) ) def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" return x * self.a + self.b class lowercase_ ( unittest.TestCase ): def lowerCamelCase_ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase_ = dummy_dataloaders() UpperCamelCase_ = ProjectConfiguration(total_limit=1 , project_dir=_a , automatic_checkpoint_naming=_a ) # Train baseline UpperCamelCase_ = Accelerator(project_config=_a ) UpperCamelCase_ = accelerator.prepare( _a , _a , _a , _a ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def lowerCamelCase_ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase_ = dummy_dataloaders() # Train baseline UpperCamelCase_ = Accelerator() UpperCamelCase_ = accelerator.prepare( _a , _a , _a , _a ) # Save initial UpperCamelCase_ = os.path.join(_a , """initial""" ) accelerator.save_state(_a ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() UpperCamelCase_ = train(3 , _a , _a , _a , _a ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() # Train partially set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase_ = dummy_dataloaders() UpperCamelCase_ = Accelerator() UpperCamelCase_ = accelerator.prepare( _a , _a , _a , _a ) accelerator.load_state(_a ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) UpperCamelCase_ = train(2 , _a , _a , _a , _a ) # Save everything UpperCamelCase_ = os.path.join(_a , """checkpoint""" ) accelerator.save_state(_a ) # Load everything back in and make sure all states work accelerator.load_state(_a ) test_rands += train(1 , _a , _a , _a , _a ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) def lowerCamelCase_ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase_ = dummy_dataloaders() UpperCamelCase_ = ProjectConfiguration(automatic_checkpoint_naming=_a ) # Train baseline UpperCamelCase_ = Accelerator(project_dir=_a , project_config=_a ) UpperCamelCase_ = accelerator.prepare( _a , _a , _a , _a ) # Save initial accelerator.save_state() (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() UpperCamelCase_ = train(3 , _a , _a , _a , _a ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() # Train partially set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase_ = dummy_dataloaders() UpperCamelCase_ = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=_a ) UpperCamelCase_ = Accelerator(project_dir=_a , project_config=_a ) UpperCamelCase_ = accelerator.prepare( _a , _a , _a , _a ) accelerator.load_state(os.path.join(_a , """checkpoints""" , """checkpoint_0""" ) ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) UpperCamelCase_ = train(2 , _a , _a , _a , _a ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(_a , """checkpoints""" , """checkpoint_1""" ) ) test_rands += train(1 , _a , _a , _a , _a ) (UpperCamelCase_) = model.a.item(), model.b.item() UpperCamelCase_ = optimizer.state_dict() self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = torch.tensor([1, 2, 3] ) UpperCamelCase_ = torch.tensor([2, 3, 4] ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(net.parameters() ) UpperCamelCase_ = Accelerator() with self.assertRaises(_a ) as ve: accelerator.register_for_checkpointing(_a , _a , _a , _a ) UpperCamelCase_ = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def lowerCamelCase_ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase_ = torch.optim.lr_scheduler.StepLR(_a , step_size=1 , gamma=0.99 ) UpperCamelCase_ = dummy_dataloaders() UpperCamelCase_ = ProjectConfiguration(automatic_checkpoint_naming=_a ) # Train baseline UpperCamelCase_ = Accelerator(project_dir=_a , project_config=_a ) UpperCamelCase_ = accelerator.prepare( _a , _a , _a , _a , _a ) # Save initial accelerator.save_state() UpperCamelCase_ = scheduler.state_dict() train(3 , _a , _a , _a , _a , _a ) self.assertNotEqual(_a , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(_a , """checkpoints""" , """checkpoint_0""" ) ) self.assertEqual(_a , scheduler.state_dict() ) def lowerCamelCase_ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(4_2 ) UpperCamelCase_ = DummyModel() UpperCamelCase_ = ProjectConfiguration(automatic_checkpoint_naming=_a , total_limit=2 ) # Train baseline UpperCamelCase_ = Accelerator(project_dir=_a , project_config=_a ) UpperCamelCase_ = accelerator.prepare(_a ) # Save 3 states: for _ in range(1_1 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(_a , """checkpoints""" , """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(_a , """checkpoints""" , """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(_a , """checkpoints""" , """checkpoint_10""" ) ) ) @require_cuda def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = ['torchrun', f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] execute_subprocess_async(_a , env=os.environ.copy() ) if __name__ == "__main__": _A = '''/tmp/accelerate/state_checkpointing''' _A = DummyModel() _A = torch.optim.Adam(params=model.parameters(), lr=1E-3) _A = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.9_9) _A , _A = dummy_dataloaders() _A = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline _A = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='''no''') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) _A , _A , _A , _A , _A = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) _A , _A = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: _A = group['''params'''][0].device break assert param_device.type == accelerator.device.type _A = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''cpu''') for group in optimizer.param_groups: _A = group['''params'''][0].device break assert ( param_device.type == torch.device('''cpu''').type ), F"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''on_device''') for group in optimizer.param_groups: _A = group['''params'''][0].device break assert ( param_device.type == accelerator.device.type ), F"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='''Unsupported optimizer map location passed'''): accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''invalid''') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
122
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class _a ( lowercase__ ): __a : Any = (DDIMParallelScheduler,) __a : Dict = (('''eta''', 0.0), ('''num_inference_steps''', 50)) def A ( self : List[str] , **lowercase : Tuple ): '''simple docstring''' UpperCAmelCase = { 'num_train_timesteps': 1_000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'clip_sample': True, } config.update(**_a ) return config def A ( self : Union[str, Any] , **lowercase : int ): '''simple docstring''' UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config(**_a ) UpperCAmelCase = scheduler_class(**_a ) UpperCAmelCase = 10, 0.0 UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter scheduler.set_timesteps(_a ) for t in scheduler.timesteps: UpperCAmelCase = model(_a , _a ) UpperCAmelCase = scheduler.step(_a , _a , _a , _a ).prev_sample return sample def A ( self : Union[str, Any] ): '''simple docstring''' for timesteps in [100, 500, 1_000]: self.check_over_configs(num_train_timesteps=_a ) def A ( self : Union[str, Any] ): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=_a ) UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase = scheduler_class(**_a ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps , torch.LongTensor([801, 601, 401, 201, 1] ) ) def A ( self : str ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=_a , beta_end=_a ) def A ( self : str ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_a ) def A ( self : Dict ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_a ) def A ( self : Dict ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=_a ) def A ( self : Union[str, Any] ): '''simple docstring''' for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=_a ) def A ( self : Optional[Any] ): '''simple docstring''' for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=_a ) def A ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=_a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=_a , prediction_type=_a , sample_max_value=_a , ) def A ( self : Any ): '''simple docstring''' for t in [1, 10, 49]: self.check_over_forward(time_step=_a ) def A ( self : str ): '''simple docstring''' for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 500] ): self.check_over_forward(time_step=_a , num_inference_steps=_a ) def A ( self : List[str] ): '''simple docstring''' for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=_a , eta=_a ) def A ( self : Any ): '''simple docstring''' UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**_a ) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(420 , 400 ) - 0.1_4771 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(980 , 960 ) - 0.3_2460 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 , 486 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 , 998 ) - 0.02 ) ) < 1E-5 def A ( self : List[Any] ): '''simple docstring''' UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**_a ) UpperCAmelCase = 10, 0.0 scheduler.set_timesteps(_a ) UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter UpperCAmelCase = self.dummy_sample_deter + 0.1 UpperCAmelCase = self.dummy_sample_deter - 0.1 UpperCAmelCase = samplea.shape[0] UpperCAmelCase = torch.stack([samplea, samplea, samplea] , dim=0 ) UpperCAmelCase = torch.arange(_a )[0:3, None].repeat(1 , _a ) UpperCAmelCase = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) UpperCAmelCase = scheduler.batch_step_no_noise(_a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) , _a ) UpperCAmelCase = torch.sum(torch.abs(_a ) ) UpperCAmelCase = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 1147.7904 ) < 1E-2 assert abs(result_mean.item() - 0.4982 ) < 1E-3 def A ( self : Any ): '''simple docstring''' UpperCAmelCase = self.full_loop() UpperCAmelCase = torch.sum(torch.abs(_a ) ) UpperCAmelCase = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 172.0067 ) < 1E-2 assert abs(result_mean.item() - 0.22_3967 ) < 1E-3 def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = self.full_loop(prediction_type='''v_prediction''' ) UpperCAmelCase = torch.sum(torch.abs(_a ) ) UpperCAmelCase = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 52.5302 ) < 1E-2 assert abs(result_mean.item() - 0.0684 ) < 1E-3 def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = self.full_loop(set_alpha_to_one=_a , beta_start=0.01 ) UpperCAmelCase = torch.sum(torch.abs(_a ) ) UpperCAmelCase = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 149.8295 ) < 1E-2 assert abs(result_mean.item() - 0.1951 ) < 1E-3 def A ( self : Tuple ): '''simple docstring''' UpperCAmelCase = self.full_loop(set_alpha_to_one=_a , beta_start=0.01 ) UpperCAmelCase = torch.sum(torch.abs(_a ) ) UpperCAmelCase = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 149.0784 ) < 1E-2 assert abs(result_mean.item() - 0.1941 ) < 1E-3
34
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor __UpperCamelCase : Dict = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def a_ ( _A ) -> int: """simple docstring""" if isinstance(__a , torch.Tensor ): return image elif isinstance(__a , PIL.Image.Image ): snake_case__ = [image] snake_case__ = [trans(img.convert('RGB' ) ) for img in image] snake_case__ = torch.stack(__a ) return image class __SCREAMING_SNAKE_CASE( lowercase__ ): def __init__( self: Optional[int] , UpperCamelCase: Optional[Any] , UpperCamelCase: List[Any] ) -> List[str]: super().__init__() # make sure scheduler can always be converted to DDIM snake_case__ = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=_a , scheduler=_a ) def lowerCAmelCase_ ( self: int , UpperCamelCase: int ) -> List[Any]: if strength < 0 or strength > 1: raise ValueError(F'''The value of strength should in [0.0, 1.0] but is {strength}''' ) def lowerCAmelCase_ ( self: Optional[int] , UpperCamelCase: Dict , UpperCamelCase: Tuple , UpperCamelCase: Dict ) -> Dict: snake_case__ = min(int(num_inference_steps * strength ) , _a ) snake_case__ = max(num_inference_steps - init_timestep , 0 ) snake_case__ = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase_ ( self: int , UpperCamelCase: List[str] , UpperCamelCase: Optional[Any] , UpperCamelCase: Union[str, Any] , UpperCamelCase: List[str] , UpperCamelCase: List[Any] , UpperCamelCase: Optional[Any]=None ) -> List[str]: if not isinstance(_a , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F'''`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_a )}''' ) snake_case__ = image.to(device=_a , dtype=_a ) if isinstance(_a , _a ) and len(_a ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(_a )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) snake_case__ = init_latents.shape snake_case__ = randn_tensor(_a , generator=_a , device=_a , dtype=_a ) # get latents print('add noise to latents at timestep' , _a ) snake_case__ = self.scheduler.add_noise(_a , _a , _a ) snake_case__ = init_latents return latents @torch.no_grad() def __call__( self: List[Any] , UpperCamelCase: Union[torch.FloatTensor, PIL.Image.Image] = None , UpperCamelCase: float = 0.8 , UpperCamelCase: int = 1 , UpperCamelCase: Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCamelCase: float = 0.0 , UpperCamelCase: int = 50 , UpperCamelCase: Optional[bool] = None , UpperCamelCase: Optional[str] = "pil" , UpperCamelCase: bool = True , ) -> Optional[Any]: self.check_inputs(_a ) # 2. Preprocess image snake_case__ = preprocess(_a ) # 3. set timesteps self.scheduler.set_timesteps(_a , device=self.device ) snake_case__ = self.get_timesteps(_a , _a , self.device ) snake_case__ = timesteps[:1].repeat(_a ) # 4. Prepare latent variables snake_case__ = self.prepare_latents(_a , _a , _a , self.unet.dtype , self.device , _a ) snake_case__ = latents # 5. Denoising loop for t in self.progress_bar(_a ): # 1. predict noise model_output snake_case__ = self.unet(_a , _a ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 snake_case__ = self.scheduler.step( _a , _a , _a , eta=_a , use_clipped_model_output=_a , generator=_a , ).prev_sample snake_case__ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case__ = self.numpy_to_pil(_a ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=_a )
307
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
a : Optional[int] = 8.3_1_4_4_5_9_8 def lowerCAmelCase_ (lowerCAmelCase__: float , lowerCAmelCase__: float ): """simple docstring""" if temperature < 0: raise Exception("""Temperature cannot be less than 0 K""" ) if molar_mass <= 0: raise Exception("""Molar mass cannot be less than or equal to 0 kg/mol""" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Optional[Any] = 300 a : Optional[int] = 28 a : Any = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
147
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename _snake_case : str = 'http://www.mocksite.com/file1.txt' _snake_case : Tuple = '\"text\": [\"foo\", \"foo\"]' _snake_case : int = '6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8' class _UpperCAmelCase : """simple docstring""" a_ = 2_00 a_ = {'''Content-Length''': '''100'''} a_ = {} def lowercase ( self : List[str] , **lowerCAmelCase_ : Tuple ) -> List[Any]: return [bytes(_a , 'utf-8' )] def a_ ( *lowerCAmelCase_ : Optional[Any], **lowerCAmelCase_ : Any ): return MockResponse() @pytest.mark.parametrize('urls_type', [str, list, dict] ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Any ): import requests monkeypatch.setattr(__a, 'request', __a ) __lowerCAmelCase = URL if issubclass(__a, __a ): __lowerCAmelCase = url elif issubclass(__a, __a ): __lowerCAmelCase = [url] elif issubclass(__a, __a ): __lowerCAmelCase = {'train': url} __lowerCAmelCase = 'dummy' __lowerCAmelCase = 'downloads' __lowerCAmelCase = tmp_path __lowerCAmelCase = DownloadConfig( cache_dir=os.path.join(__a, __a ), use_etag=__a, ) __lowerCAmelCase = DownloadManager(dataset_name=__a, download_config=__a ) __lowerCAmelCase = dl_manager.download(__a ) __lowerCAmelCase = urls for downloaded_paths in [downloaded_paths]: if isinstance(__a, __a ): __lowerCAmelCase = [downloaded_paths] __lowerCAmelCase = [urls] elif isinstance(__a, __a ): assert "train" in downloaded_paths.keys() __lowerCAmelCase = downloaded_paths.values() __lowerCAmelCase = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(__a, __a ): assert downloaded_path == dl_manager.downloaded_paths[input_url] __lowerCAmelCase = Path(__a ) __lowerCAmelCase = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() __lowerCAmelCase = downloaded_path.read_text() assert content == CONTENT __lowerCAmelCase = downloaded_path.with_suffix('.json' ) assert metadata_downloaded_path.exists() __lowerCAmelCase = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize('paths_type', [str, list, dict] ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[Any], lowerCAmelCase_ : Any ): __lowerCAmelCase = str(__a ) if issubclass(__a, __a ): __lowerCAmelCase = filename elif issubclass(__a, __a ): __lowerCAmelCase = [filename] elif issubclass(__a, __a ): __lowerCAmelCase = {'train': filename} __lowerCAmelCase = 'dummy' __lowerCAmelCase = xz_file.parent __lowerCAmelCase = 'extracted' __lowerCAmelCase = DownloadConfig( cache_dir=__a, use_etag=__a, ) __lowerCAmelCase = DownloadManager(dataset_name=__a, download_config=__a ) __lowerCAmelCase = dl_manager.extract(__a ) __lowerCAmelCase = paths for extracted_paths in [extracted_paths]: if isinstance(__a, __a ): __lowerCAmelCase = [extracted_paths] __lowerCAmelCase = [paths] elif isinstance(__a, __a ): assert "train" in extracted_paths.keys() __lowerCAmelCase = extracted_paths.values() __lowerCAmelCase = paths.values() assert extracted_paths for extracted_path, input_path in zip(__a, __a ): assert extracted_path == dl_manager.extracted_paths[input_path] __lowerCAmelCase = Path(__a ) __lowerCAmelCase = extracted_path.parts assert parts[-1] == hash_url_to_filename(__a, etag=__a ) assert parts[-2] == extracted_subdir assert extracted_path.exists() __lowerCAmelCase = extracted_path.read_text() __lowerCAmelCase = text_file.read_text() assert extracted_file_content == expected_file_content def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict ): assert path.endswith('.jsonl' ) for num_items, line in enumerate(__a, start=1 ): __lowerCAmelCase = json.loads(line.decode('utf-8' ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize('archive_jsonl', ['tar_jsonl_path', 'zip_jsonl_path'] ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[str] ): __lowerCAmelCase = request.getfixturevalue(__a ) __lowerCAmelCase = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(__a ), start=1 ): _test_jsonl(__a, __a ) assert num_jsonl == 2 @pytest.mark.parametrize('archive_nested_jsonl', ['tar_nested_jsonl_path', 'zip_nested_jsonl_path'] ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : int ): __lowerCAmelCase = request.getfixturevalue(__a ) __lowerCAmelCase = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(__a ), start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(__a ), start=1 ): _test_jsonl(__a, __a ) assert num_tar == 1 assert num_jsonl == 2 def a_ ( lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(__a ), start=1 ): assert os.path.basename(__a ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
284
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) class _a ( lowercase__): _a : str = '''timm_backbone''' def __init__( self : int , _SCREAMING_SNAKE_CASE : Union[str, Any]=None , _SCREAMING_SNAKE_CASE : Union[str, Any]=3 , _SCREAMING_SNAKE_CASE : str=True , _SCREAMING_SNAKE_CASE : str=True , _SCREAMING_SNAKE_CASE : List[str]=None , **_SCREAMING_SNAKE_CASE : Optional[int] , )-> List[Any]: super().__init__(**_a ) lowerCAmelCase__ : Union[str, Any] = backbone lowerCAmelCase__ : Union[str, Any] = num_channels lowerCAmelCase__ : List[Any] = features_only lowerCAmelCase__ : Tuple = use_pretrained_backbone lowerCAmelCase__ : Any = True lowerCAmelCase__ : Optional[int] = out_indices if out_indices is not None else (-1,)
131
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''ctc_proj''', '''mask_emb''': '''masked_spec_embed''', } lowerCAmelCase__ = [ '''ctc_proj''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def a__ ( SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' for attribute in key.split("." ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models lowerCAmelCase : Any = 'lm_head' lowerCAmelCase : Any = getattr(__a , __a ) if weight_type is not None: lowerCAmelCase : List[Any] = getattr(__a , __a ).shape else: lowerCAmelCase : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": lowerCAmelCase : List[Any] = value elif weight_type == "weight_g": lowerCAmelCase : Union[str, Any] = value elif weight_type == "weight_v": lowerCAmelCase : Dict = value elif weight_type == "bias": lowerCAmelCase : Optional[Any] = value else: lowerCAmelCase : Any = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' lowerCAmelCase : int = [] lowerCAmelCase : List[str] = fairseq_model.state_dict() lowerCAmelCase : List[Any] = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): lowerCAmelCase : Any = False if "conv_layers" in name: load_conv_layer( __a , __a , __a , __a , hf_model.config.feat_extract_norm == "group" , ) lowerCAmelCase : List[str] = True else: for key, mapped_key in MAPPING.items(): lowerCAmelCase : Optional[int] = 'unispeech.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: lowerCAmelCase : Union[str, Any] = True if "*" in mapped_key: lowerCAmelCase : Dict = name.split(__a )[0].split("." )[-2] lowerCAmelCase : Optional[int] = mapped_key.replace("*" , __a ) if "weight_g" in name: lowerCAmelCase : List[Any] = 'weight_g' elif "weight_v" in name: lowerCAmelCase : str = 'weight_v' elif "bias" in name: lowerCAmelCase : Dict = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj lowerCAmelCase : str = 'weight' else: lowerCAmelCase : List[str] = None set_recursively(__a , __a , __a , __a , __a , __a ) continue if not is_used: unused_weights.append(__a ) logger.warning(f"""Unused weights: {unused_weights}""" ) def a__ ( SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' lowerCAmelCase : Any = full_name.split("conv_layers." )[-1] lowerCAmelCase : int = name.split("." ) lowerCAmelCase : Optional[Any] = int(items[0] ) lowerCAmelCase : Union[str, Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) lowerCAmelCase : int = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) lowerCAmelCase : Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) lowerCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) lowerCAmelCase : Optional[Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__a ) @torch.no_grad() def a__ ( SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Any=None , SCREAMING_SNAKE_CASE : Optional[Any]=None , SCREAMING_SNAKE_CASE : Union[str, Any]=True ): '''simple docstring''' if config_path is not None: lowerCAmelCase : Optional[Any] = UniSpeechConfig.from_pretrained(__a ) else: lowerCAmelCase : str = UniSpeechConfig() if is_finetuned: if dict_path: lowerCAmelCase : int = Dictionary.load_from_json(__a ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCAmelCase : Optional[int] = target_dict.pad_index lowerCAmelCase : Union[str, Any] = target_dict.bos_index lowerCAmelCase : List[str] = target_dict.eos_index lowerCAmelCase : List[str] = len(target_dict.symbols ) lowerCAmelCase : Optional[Any] = os.path.join(__a , "vocab.json" ) if not os.path.isdir(__a ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(__a ) ) return os.makedirs(__a , exist_ok=__a ) lowerCAmelCase : Optional[Any] = target_dict.indices # fairseq has the <pad> and <s> switched lowerCAmelCase : Optional[Any] = 4_2 lowerCAmelCase : Any = 4_3 with open(__a , "w" , encoding="utf-8" ) as vocab_handle: json.dump(__a , __a ) lowerCAmelCase : Tuple = WavaVecaPhonemeCTCTokenizer( __a , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=__a , ) lowerCAmelCase : str = True if config.feat_extract_norm == 'layer' else False lowerCAmelCase : Tuple = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=__a , return_attention_mask=__a , ) lowerCAmelCase : int = WavaVecaProcessor(feature_extractor=__a , tokenizer=__a ) processor.save_pretrained(__a ) lowerCAmelCase : str = UniSpeechForCTC(__a ) else: lowerCAmelCase : str = UniSpeechForPreTraining(__a ) if is_finetuned: lowerCAmelCase : str = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] ), "w2v_path": checkpoint_path} ) else: lowerCAmelCase : Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) lowerCAmelCase : Union[str, Any] = model[0].eval() recursively_load_weights(__a , __a , __a ) hf_unispeech.save_pretrained(__a ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) lowerCAmelCase__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
108
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( lowercase__ ): '''simple docstring''' lowerCamelCase__ =['''image_processor''', '''tokenizer'''] lowerCamelCase__ ='''LayoutLMv2ImageProcessor''' lowerCamelCase__ =('''LayoutXLMTokenizer''', '''LayoutXLMTokenizerFast''') def __init__(self , a_=None , a_=None , **a_ ): '''simple docstring''' if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _a , ) __snake_case : Union[str, Any] = kwargs.pop('''feature_extractor''' ) __snake_case : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_a , _a ) def __call__(self , a_ , a_ = None , a_ = None , a_ = None , a_ = None , a_ = True , a_ = False , a_ = None , a_ = None , a_ = 0 , a_ = None , a_ = None , a_ = None , a_ = False , a_ = False , a_ = False , a_ = False , a_ = True , a_ = None , **a_ , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes ''' '''if you initialized the image processor with apply_ocr set to True.''' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''' ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('''You cannot return overflowing tokens without returning the offsets mapping.''' ) # first, apply the image processor __snake_case : List[Any] = self.image_processor(images=_a , return_tensors=_a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(_a , _a ): __snake_case : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) __snake_case : Tuple = features['words'] __snake_case : Any = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=_a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , stride=_a , pad_to_multiple_of=_a , return_token_type_ids=_a , return_attention_mask=_a , return_overflowing_tokens=_a , return_special_tokens_mask=_a , return_offsets_mapping=_a , return_length=_a , verbose=_a , return_tensors=_a , **_a , ) # add pixel values __snake_case : List[Any] = features.pop('''pixel_values''' ) if return_overflowing_tokens is True: __snake_case : Optional[Any] = self.get_overflowing_images(_a , encoded_inputs['''overflow_to_sample_mapping'''] ) __snake_case : Tuple = images return encoded_inputs def SCREAMING_SNAKE_CASE (self , a_ , a_ ): '''simple docstring''' __snake_case : List[str] = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(_a ) != len(_a ): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' f""" {len(_a )} and {len(_a )}""" ) return images_with_overflow def SCREAMING_SNAKE_CASE (self , *a_ , **a_ ): '''simple docstring''' return self.tokenizer.batch_decode(*_a , **_a ) def SCREAMING_SNAKE_CASE (self , *a_ , **a_ ): '''simple docstring''' return self.tokenizer.decode(*_a , **_a ) @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "image"] @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _a , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _a , ) return self.image_processor
102
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def __magic_name__ ( __snake_case : int ) -> Tuple: random.seed(__a ) np.random.seed(__a ) torch.manual_seed(__a ) torch.cuda.manual_seed_all(__a ) # ^^ safe to call this function even if cuda is not available class a__ : def __init__( self , _a , _a = 0.9_9_9_9 , _a = 0.0 , _a = 0 , _a = False , _a = 1.0 , _a = 2 / 3 , _a = None , _a = None , **_a , ): if isinstance(_a , torch.nn.Module ): lowercase : str = ( 'Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ' 'Please pass the parameters of the module instead.' ) deprecate( "passing a `torch.nn.Module` to `ExponentialMovingAverage`" , "1.0.0" , _a , standard_warn=_a , ) lowercase : Optional[int] = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility lowercase : List[str] = True if kwargs.get("max_value" , _a ) is not None: lowercase : Optional[int] = 'The `max_value` argument is deprecated. Please use `decay` instead.' deprecate("max_value" , "1.0.0" , _a , standard_warn=_a ) lowercase : Union[str, Any] = kwargs['max_value'] if kwargs.get("min_value" , _a ) is not None: lowercase : List[str] = 'The `min_value` argument is deprecated. Please use `min_decay` instead.' deprecate("min_value" , "1.0.0" , _a , standard_warn=_a ) lowercase : Optional[int] = kwargs['min_value'] lowercase : Tuple = list(_a ) lowercase : List[str] = [p.clone().detach() for p in parameters] if kwargs.get("device" , _a ) is not None: lowercase : str = 'The `device` argument is deprecated. Please use `to` instead.' deprecate("device" , "1.0.0" , _a , standard_warn=_a ) self.to(device=kwargs["device"] ) lowercase : str = None lowercase : Optional[Any] = decay lowercase : Union[str, Any] = min_decay lowercase : Tuple = update_after_step lowercase : Union[str, Any] = use_ema_warmup lowercase : str = inv_gamma lowercase : int = power lowercase : Tuple = 0 lowercase : Any = None # set in `step()` lowercase : List[str] = model_cls lowercase : Tuple = model_config @classmethod def __magic_name__ ( cls , _a , _a ): lowercase : Union[str, Any] = model_cls.load_config(_a , return_unused_kwargs=_a ) lowercase : Dict = model_cls.from_pretrained(_a ) lowercase : List[str] = cls(model.parameters() , model_cls=_a , model_config=model.config ) ema_model.load_state_dict(_a ) return ema_model def __magic_name__ ( self , _a ): if self.model_cls is None: raise ValueError("`save_pretrained` can only be used if `model_cls` was defined at __init__." ) if self.model_config is None: raise ValueError("`save_pretrained` can only be used if `model_config` was defined at __init__." ) lowercase : str = self.model_cls.from_config(self.model_config ) lowercase : List[str] = self.state_dict() state_dict.pop("shadow_params" , _a ) model.register_to_config(**_a ) self.copy_to(model.parameters() ) model.save_pretrained(_a ) def __magic_name__ ( self , _a ): lowercase : Tuple = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: lowercase : Optional[int] = 1 - (1 + step / self.inv_gamma) ** -self.power else: lowercase : List[Any] = (1 + step) / (10 + step) lowercase : Tuple = min(_a , self.decay ) # make sure decay is not smaller than min_decay lowercase : str = max(_a , self.min_decay ) return cur_decay_value @torch.no_grad() def __magic_name__ ( self , _a ): if isinstance(_a , torch.nn.Module ): lowercase : str = ( 'Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ' 'Please pass the parameters of the module instead.' ) deprecate( "passing a `torch.nn.Module` to `ExponentialMovingAverage.step`" , "1.0.0" , _a , standard_warn=_a , ) lowercase : Tuple = parameters.parameters() lowercase : Optional[Any] = list(_a ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. lowercase : int = self.get_decay(self.optimization_step ) lowercase : Dict = decay lowercase : int = 1 - decay lowercase : Union[str, Any] = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , _a ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): lowercase : Union[str, Any] = deepspeed.zero.GatheredParameters(_a , modifier_rank=_a ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(_a ) def __magic_name__ ( self , _a ): lowercase : str = list(_a ) for s_param, param in zip(self.shadow_params , _a ): param.data.copy_(s_param.to(param.device ).data ) def __magic_name__ ( self , _a=None , _a=None ): lowercase : str = [ p.to(device=_a , dtype=_a ) if p.is_floating_point() else p.to(device=_a ) for p in self.shadow_params ] def __magic_name__ ( self ): return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def __magic_name__ ( self , _a ): lowercase : List[Any] = [param.detach().cpu().clone() for param in parameters] def __magic_name__ ( self , _a ): if self.temp_stored_params is None: raise RuntimeError("This ExponentialMovingAverage has no `store()`ed weights " "to `restore()`" ) for c_param, param in zip(self.temp_stored_params , _a ): param.data.copy_(c_param.data ) # Better memory-wise. lowercase : Union[str, Any] = None def __magic_name__ ( self , _a ): lowercase : List[Any] = copy.deepcopy(_a ) lowercase : Optional[int] = state_dict.get("decay" , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError("Decay must be between 0 and 1" ) lowercase : List[Any] = state_dict.get("min_decay" , self.min_decay ) if not isinstance(self.min_decay , _a ): raise ValueError("Invalid min_decay" ) lowercase : str = state_dict.get("optimization_step" , self.optimization_step ) if not isinstance(self.optimization_step , _a ): raise ValueError("Invalid optimization_step" ) lowercase : Optional[int] = state_dict.get("update_after_step" , self.update_after_step ) if not isinstance(self.update_after_step , _a ): raise ValueError("Invalid update_after_step" ) lowercase : str = state_dict.get("use_ema_warmup" , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , _a ): raise ValueError("Invalid use_ema_warmup" ) lowercase : int = state_dict.get("inv_gamma" , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError("Invalid inv_gamma" ) lowercase : Optional[Any] = state_dict.get("power" , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError("Invalid power" ) lowercase : Union[str, Any] = state_dict.get("shadow_params" , _a ) if shadow_params is not None: lowercase : Union[str, Any] = shadow_params if not isinstance(self.shadow_params , _a ): raise ValueError("shadow_params must be a list" ) if not all(isinstance(_a , torch.Tensor ) for p in self.shadow_params ): raise ValueError("shadow_params must all be Tensors" )
202
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ : Any = { """configuration_xlm_roberta""": [ """XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaConfig""", """XLMRobertaOnnxConfig""", ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Union[str, Any] = ["""XLMRobertaTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Union[str, Any] = ["""XLMRobertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Optional[Any] = [ """XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaForCausalLM""", """XLMRobertaForMaskedLM""", """XLMRobertaForMultipleChoice""", """XLMRobertaForQuestionAnswering""", """XLMRobertaForSequenceClassification""", """XLMRobertaForTokenClassification""", """XLMRobertaModel""", """XLMRobertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Union[str, Any] = [ """TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFXLMRobertaForCausalLM""", """TFXLMRobertaForMaskedLM""", """TFXLMRobertaForMultipleChoice""", """TFXLMRobertaForQuestionAnswering""", """TFXLMRobertaForSequenceClassification""", """TFXLMRobertaForTokenClassification""", """TFXLMRobertaModel""", """TFXLMRobertaPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : List[str] = [ """FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """FlaxXLMRobertaForMaskedLM""", """FlaxXLMRobertaForCausalLM""", """FlaxXLMRobertaForMultipleChoice""", """FlaxXLMRobertaForQuestionAnswering""", """FlaxXLMRobertaForSequenceClassification""", """FlaxXLMRobertaForTokenClassification""", """FlaxXLMRobertaModel""", """FlaxXLMRobertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys lowercase__ : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
224
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
'''simple docstring''' from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : "DiagonalGaussianDistribution" class a__( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : str = True @register_to_config def __init__( self , __lowerCAmelCase = 3 , __lowerCAmelCase = 3 , __lowerCAmelCase = ("DownEncoderBlock2D",) , __lowerCAmelCase = ("UpDecoderBlock2D",) , __lowerCAmelCase = (64,) , __lowerCAmelCase = 1 , __lowerCAmelCase = "silu" , __lowerCAmelCase = 4 , __lowerCAmelCase = 32 , __lowerCAmelCase = 32 , __lowerCAmelCase = 0.18215 , ): """simple docstring""" super().__init__() # pass init params to Encoder lowerCAmelCase = Encoder( in_channels=__lowerCAmelCase , out_channels=__lowerCAmelCase , down_block_types=__lowerCAmelCase , block_out_channels=__lowerCAmelCase , layers_per_block=__lowerCAmelCase , act_fn=__lowerCAmelCase , norm_num_groups=__lowerCAmelCase , double_z=__lowerCAmelCase , ) # pass init params to Decoder lowerCAmelCase = Decoder( in_channels=__lowerCAmelCase , out_channels=__lowerCAmelCase , up_block_types=__lowerCAmelCase , block_out_channels=__lowerCAmelCase , layers_per_block=__lowerCAmelCase , norm_num_groups=__lowerCAmelCase , act_fn=__lowerCAmelCase , ) lowerCAmelCase = nn.Convad(2 * latent_channels , 2 * latent_channels , 1) lowerCAmelCase = nn.Convad(__lowerCAmelCase , __lowerCAmelCase , 1) lowerCAmelCase = False lowerCAmelCase = False # only relevant if vae tiling is enabled lowerCAmelCase = self.config.sample_size lowerCAmelCase = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple)) else self.config.sample_size ) lowerCAmelCase = int(sample_size / (2 ** (len(self.config.block_out_channels) - 1))) lowerCAmelCase = 0.25 def a_ ( self , __lowerCAmelCase , __lowerCAmelCase=False): """simple docstring""" if isinstance(__lowerCAmelCase , (Encoder, Decoder)): lowerCAmelCase = value def a_ ( self , __lowerCAmelCase = True): """simple docstring""" lowerCAmelCase = use_tiling def a_ ( self): """simple docstring""" self.enable_tiling(__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = True def a_ ( self): """simple docstring""" lowerCAmelCase = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def a_ ( self): """simple docstring""" lowerCAmelCase = {} def fn_recursive_add_processors(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): if hasattr(__lowerCAmelCase , """set_processor"""): lowerCAmelCase = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f"{name}.{sub_name}" , __lowerCAmelCase , __lowerCAmelCase) return processors for name, module in self.named_children(): fn_recursive_add_processors(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) return processors def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = len(self.attn_processors.keys()) if isinstance(__lowerCAmelCase , __lowerCAmelCase) and len(__lowerCAmelCase) != count: raise ValueError( f"A dict of processors was passed, but the number of processors {len(__lowerCAmelCase)} does not match the" f" number of attention layers: {count}. Please make sure to pass {count} processor classes.") def fn_recursive_attn_processor(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): if hasattr(__lowerCAmelCase , """set_processor"""): if not isinstance(__lowerCAmelCase , __lowerCAmelCase): module.set_processor(__lowerCAmelCase) else: module.set_processor(processor.pop(f"{name}.processor")) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f"{name}.{sub_name}" , __lowerCAmelCase , __lowerCAmelCase) for name, module in self.named_children(): fn_recursive_attn_processor(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) def a_ ( self): """simple docstring""" self.set_attn_processor(AttnProcessor()) @apply_forward_hook def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = True): """simple docstring""" if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(__lowerCAmelCase , return_dict=__lowerCAmelCase) if self.use_slicing and x.shape[0] > 1: lowerCAmelCase = [self.encoder(__lowerCAmelCase) for x_slice in x.split(1)] lowerCAmelCase = torch.cat(__lowerCAmelCase) else: lowerCAmelCase = self.encoder(__lowerCAmelCase) lowerCAmelCase = self.quant_conv(__lowerCAmelCase) lowerCAmelCase = DiagonalGaussianDistribution(__lowerCAmelCase) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = True): """simple docstring""" if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(__lowerCAmelCase , return_dict=__lowerCAmelCase) lowerCAmelCase = self.post_quant_conv(__lowerCAmelCase) lowerCAmelCase = self.decoder(__lowerCAmelCase) if not return_dict: return (dec,) return DecoderOutput(sample=__lowerCAmelCase) @apply_forward_hook def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = True): """simple docstring""" if self.use_slicing and z.shape[0] > 1: lowerCAmelCase = [self._decode(__lowerCAmelCase).sample for z_slice in z.split(1)] lowerCAmelCase = torch.cat(__lowerCAmelCase) else: lowerCAmelCase = self._decode(__lowerCAmelCase).sample if not return_dict: return (decoded,) return DecoderOutput(sample=__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = min(a.shape[2] , b.shape[2] , __lowerCAmelCase) for y in range(__lowerCAmelCase): lowerCAmelCase = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = min(a.shape[3] , b.shape[3] , __lowerCAmelCase) for x in range(__lowerCAmelCase): lowerCAmelCase = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = True): """simple docstring""" lowerCAmelCase = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor)) lowerCAmelCase = int(self.tile_latent_min_size * self.tile_overlap_factor) lowerCAmelCase = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. lowerCAmelCase = [] for i in range(0 , x.shape[2] , __lowerCAmelCase): lowerCAmelCase = [] for j in range(0 , x.shape[3] , __lowerCAmelCase): lowerCAmelCase = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] lowerCAmelCase = self.encoder(__lowerCAmelCase) lowerCAmelCase = self.quant_conv(__lowerCAmelCase) row.append(__lowerCAmelCase) rows.append(__lowerCAmelCase) lowerCAmelCase = [] for i, row in enumerate(__lowerCAmelCase): lowerCAmelCase = [] for j, tile in enumerate(__lowerCAmelCase): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: lowerCAmelCase = self.blend_v(rows[i - 1][j] , __lowerCAmelCase , __lowerCAmelCase) if j > 0: lowerCAmelCase = self.blend_h(row[j - 1] , __lowerCAmelCase , __lowerCAmelCase) result_row.append(tile[:, :, :row_limit, :row_limit]) result_rows.append(torch.cat(__lowerCAmelCase , dim=3)) lowerCAmelCase = torch.cat(__lowerCAmelCase , dim=2) lowerCAmelCase = DiagonalGaussianDistribution(__lowerCAmelCase) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = True): """simple docstring""" lowerCAmelCase = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor)) lowerCAmelCase = int(self.tile_sample_min_size * self.tile_overlap_factor) lowerCAmelCase = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. lowerCAmelCase = [] for i in range(0 , z.shape[2] , __lowerCAmelCase): lowerCAmelCase = [] for j in range(0 , z.shape[3] , __lowerCAmelCase): lowerCAmelCase = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] lowerCAmelCase = self.post_quant_conv(__lowerCAmelCase) lowerCAmelCase = self.decoder(__lowerCAmelCase) row.append(__lowerCAmelCase) rows.append(__lowerCAmelCase) lowerCAmelCase = [] for i, row in enumerate(__lowerCAmelCase): lowerCAmelCase = [] for j, tile in enumerate(__lowerCAmelCase): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: lowerCAmelCase = self.blend_v(rows[i - 1][j] , __lowerCAmelCase , __lowerCAmelCase) if j > 0: lowerCAmelCase = self.blend_h(row[j - 1] , __lowerCAmelCase , __lowerCAmelCase) result_row.append(tile[:, :, :row_limit, :row_limit]) result_rows.append(torch.cat(__lowerCAmelCase , dim=3)) lowerCAmelCase = torch.cat(__lowerCAmelCase , dim=2) if not return_dict: return (dec,) return DecoderOutput(sample=__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = False , __lowerCAmelCase = True , __lowerCAmelCase = None , ): """simple docstring""" lowerCAmelCase = sample lowerCAmelCase = self.encode(__lowerCAmelCase).latent_dist if sample_posterior: lowerCAmelCase = posterior.sample(generator=__lowerCAmelCase) else: lowerCAmelCase = posterior.mode() lowerCAmelCase = self.decode(__lowerCAmelCase).sample if not return_dict: return (dec,) return DecoderOutput(sample=__lowerCAmelCase)
272
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __lowercase = logging.get_logger(__name__) class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = ['''input_features'''] def __init__( self , __lowerCAmelCase=80 , __lowerCAmelCase=16000 , __lowerCAmelCase=160 , __lowerCAmelCase=30 , __lowerCAmelCase=400 , __lowerCAmelCase=0.0 , __lowerCAmelCase=False , **__lowerCAmelCase , ): """simple docstring""" super().__init__( feature_size=__lowerCAmelCase , sampling_rate=__lowerCAmelCase , padding_value=__lowerCAmelCase , return_attention_mask=__lowerCAmelCase , **__lowerCAmelCase , ) lowerCAmelCase = n_fft lowerCAmelCase = hop_length lowerCAmelCase = chunk_length lowerCAmelCase = chunk_length * sampling_rate lowerCAmelCase = self.n_samples // hop_length lowerCAmelCase = sampling_rate lowerCAmelCase = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__lowerCAmelCase , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=__lowerCAmelCase , norm="""slaney""" , mel_scale="""slaney""" , ) def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = spectrogram( __lowerCAmelCase , window_function(self.n_fft , """hann""") , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="""log10""" , ) lowerCAmelCase = log_spec[:, :-1] lowerCAmelCase = np.maximum(__lowerCAmelCase , log_spec.max() - 8.0) lowerCAmelCase = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def a_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 0.0): """simple docstring""" if attention_mask is not None: lowerCAmelCase = np.array(__lowerCAmelCase , np.intaa) lowerCAmelCase = [] for vector, length in zip(__lowerCAmelCase , attention_mask.sum(-1)): lowerCAmelCase = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7) if length < normed_slice.shape[0]: lowerCAmelCase = padding_value normed_input_values.append(__lowerCAmelCase) else: lowerCAmelCase = [(x - x.mean()) / np.sqrt(x.var() + 1E-7) for x in input_values] return normed_input_values def __call__( self , __lowerCAmelCase , __lowerCAmelCase = True , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = "max_length" , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , **__lowerCAmelCase , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" f" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" f" was sampled with {self.sampling_rate} and not {sampling_rate}.") else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""") lowerCAmelCase = isinstance(__lowerCAmelCase , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}") lowerCAmelCase = is_batched_numpy or ( isinstance(__lowerCAmelCase , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: lowerCAmelCase = [np.asarray([speech] , dtype=np.floataa).T for speech in raw_speech] elif not is_batched and not isinstance(__lowerCAmelCase , np.ndarray): lowerCAmelCase = np.asarray(__lowerCAmelCase , dtype=np.floataa) elif isinstance(__lowerCAmelCase , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): lowerCAmelCase = raw_speech.astype(np.floataa) # always return batch if not is_batched: lowerCAmelCase = [np.asarray([raw_speech]).T] lowerCAmelCase = BatchFeature({"""input_features""": raw_speech}) # convert into correct format for padding lowerCAmelCase = self.pad( __lowerCAmelCase , padding=__lowerCAmelCase , max_length=max_length if max_length else self.n_samples , truncation=__lowerCAmelCase , pad_to_multiple_of=__lowerCAmelCase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: lowerCAmelCase = self.zero_mean_unit_var_norm( padded_inputs["""input_features"""] , attention_mask=padded_inputs["""attention_mask"""] , padding_value=self.padding_value , ) lowerCAmelCase = np.stack(padded_inputs["""input_features"""] , axis=0) # make sure list is in array format lowerCAmelCase = padded_inputs.get("""input_features""").transpose(2 , 0 , 1) lowerCAmelCase = [self._np_extract_fbank_features(__lowerCAmelCase) for waveform in input_features[0]] if isinstance(input_features[0] , __lowerCAmelCase): lowerCAmelCase = [np.asarray(__lowerCAmelCase , dtype=np.floataa) for feature in input_features] else: lowerCAmelCase = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) lowerCAmelCase = padded_inputs["""attention_mask"""][:, :: self.hop_length] if return_tensors is not None: lowerCAmelCase = padded_inputs.convert_to_tensors(__lowerCAmelCase) return padded_inputs def a_ ( self): """simple docstring""" lowerCAmelCase = copy.deepcopy(self.__dict__) lowerCAmelCase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
272
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class a__: '''simple docstring''' UpperCAmelCase_ : Tuple = BlenderbotConfig UpperCAmelCase_ : str = {} UpperCAmelCase_ : Optional[Any] = '''gelu''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=7 , __lowerCAmelCase=True , __lowerCAmelCase=False , __lowerCAmelCase=99 , __lowerCAmelCase=32 , __lowerCAmelCase=2 , __lowerCAmelCase=4 , __lowerCAmelCase=37 , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=20 , __lowerCAmelCase=2 , __lowerCAmelCase=1 , __lowerCAmelCase=0 , ): """simple docstring""" lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = eos_token_id lowerCAmelCase = pad_token_id lowerCAmelCase = bos_token_id def a_ ( self): """simple docstring""" lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowerCAmelCase = prepare_blenderbot_inputs_dict(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) return config, inputs_dict def a_ ( self , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = TFBlenderbotModel(config=__lowerCAmelCase).get_decoder() lowerCAmelCase = inputs_dict["""input_ids"""] lowerCAmelCase = input_ids[:1, :] lowerCAmelCase = inputs_dict["""attention_mask"""][:1, :] lowerCAmelCase = inputs_dict["""head_mask"""] lowerCAmelCase = 1 # first forward pass lowerCAmelCase = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , head_mask=__lowerCAmelCase , use_cache=__lowerCAmelCase) lowerCAmelCase , lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size) lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2) , tf.inta) # append to next input_ids and lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1) lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1) lowerCAmelCase = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase)[0] lowerCAmelCase = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , past_key_values=__lowerCAmelCase)[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1]) # select random slice lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1])) lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__lowerCAmelCase , __lowerCAmelCase , rtol=1E-3) def snake_case__ ( _A: Any , _A: Optional[int] , _A: Optional[int] , _A: List[str]=None , _A: str=None , _A: Union[str, Any]=None , _A: Tuple=None , _A: Optional[Any]=None , ) -> Optional[Any]: '''simple docstring''' if attention_mask is None: lowerCAmelCase = tf.cast(tf.math.not_equal(_A , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class a__( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : str = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () UpperCAmelCase_ : Any = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () UpperCAmelCase_ : Optional[int] = ( { '''conversational''': TFBlenderbotForConditionalGeneration, '''feature-extraction''': TFBlenderbotModel, '''summarization''': TFBlenderbotForConditionalGeneration, '''text2text-generation''': TFBlenderbotForConditionalGeneration, '''translation''': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = False UpperCAmelCase_ : List[str] = False def a_ ( self): """simple docstring""" lowerCAmelCase = TFBlenderbotModelTester(self) lowerCAmelCase = ConfigTester(self , config_class=__lowerCAmelCase) def a_ ( self): """simple docstring""" self.config_tester.run_common_tests() def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__lowerCAmelCase) @require_tokenizers @require_tf class a__( unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : List[str] = ['''My friends are cool but they eat too many carbs.'''] UpperCAmelCase_ : Union[str, Any] = '''facebook/blenderbot-400M-distill''' @cached_property def a_ ( self): """simple docstring""" return BlenderbotTokenizer.from_pretrained(self.model_name) @cached_property def a_ ( self): """simple docstring""" lowerCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model @slow def a_ ( self): """simple docstring""" lowerCAmelCase = self.tokenizer(self.src_text , return_tensors="""tf""") lowerCAmelCase = self.model.generate( model_inputs.input_ids , ) lowerCAmelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__lowerCAmelCase)[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
272
'''simple docstring''' from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig __lowercase = logging.get_logger(__name__) __lowercase = '''T5Config''' class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : List[str] = '''mt5''' UpperCAmelCase_ : Tuple = MTaConfig class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : List[Any] = '''mt5''' UpperCAmelCase_ : int = MTaConfig class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : Tuple = '''mt5''' UpperCAmelCase_ : Union[str, Any] = MTaConfig
272
1
'''simple docstring''' import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def snake_case__ ( _A: Optional[int] ) -> Tuple: '''simple docstring''' return EnvironmentCommand() def snake_case__ ( _A: List[Any] ) -> int: '''simple docstring''' return EnvironmentCommand(args.accelerate_config_file ) class a__( lowerCAmelCase__ ): '''simple docstring''' @staticmethod def a_ ( __lowerCAmelCase): """simple docstring""" lowerCAmelCase = parser.add_parser("""env""") download_parser.set_defaults(func=__lowerCAmelCase) download_parser.add_argument( """--accelerate-config_file""" , default=__lowerCAmelCase , help="""The accelerate config file to use for the default values in the launching script.""" , ) download_parser.set_defaults(func=__lowerCAmelCase) def __init__( self , __lowerCAmelCase , *__lowerCAmelCase): """simple docstring""" lowerCAmelCase = accelerate_config_file def a_ ( self): """simple docstring""" lowerCAmelCase = """not installed""" if is_safetensors_available(): import safetensors lowerCAmelCase = safetensors.__version__ elif importlib.util.find_spec("""safetensors""") is not None: import safetensors lowerCAmelCase = f"{safetensors.__version__} but is ignored because of PyTorch version too old." lowerCAmelCase = """not installed""" lowerCAmelCase = lowerCAmelCase = """not found""" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file lowerCAmelCase = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(__lowerCAmelCase): lowerCAmelCase = load_config_from_file(self._accelerate_config_file).to_dict() lowerCAmelCase = ( """\n""".join([f"\t- {prop}: {val}" for prop, val in accelerate_config.items()]) if isinstance(__lowerCAmelCase , __lowerCAmelCase) else f"\t{accelerate_config}" ) lowerCAmelCase = """not installed""" lowerCAmelCase = """NA""" if is_torch_available(): import torch lowerCAmelCase = torch.__version__ lowerCAmelCase = torch.cuda.is_available() lowerCAmelCase = """not installed""" lowerCAmelCase = """NA""" if is_tf_available(): import tensorflow as tf lowerCAmelCase = tf.__version__ try: # deprecated in v2.1 lowerCAmelCase = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool lowerCAmelCase = bool(tf.config.list_physical_devices("""GPU""")) lowerCAmelCase = """not installed""" lowerCAmelCase = """not installed""" lowerCAmelCase = """not installed""" lowerCAmelCase = """NA""" if is_flax_available(): import flax import jax import jaxlib lowerCAmelCase = flax.__version__ lowerCAmelCase = jax.__version__ lowerCAmelCase = jaxlib.__version__ lowerCAmelCase = jax.lib.xla_bridge.get_backend().platform lowerCAmelCase = { """`transformers` version""": version, """Platform""": platform.platform(), """Python version""": platform.python_version(), """Huggingface_hub version""": huggingface_hub.__version__, """Safetensors version""": f"{safetensors_version}", """Accelerate version""": f"{accelerate_version}", """Accelerate config""": f"{accelerate_config_str}", """PyTorch version (GPU?)""": f"{pt_version} ({pt_cuda_available})", """Tensorflow version (GPU?)""": f"{tf_version} ({tf_cuda_available})", """Flax version (CPU?/GPU?/TPU?)""": f"{flax_version} ({jax_backend})", """Jax version""": f"{jax_version}", """JaxLib version""": f"{jaxlib_version}", """Using GPU in script?""": """<fill in>""", """Using distributed or parallel set-up in script?""": """<fill in>""", } print("""\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n""") print(self.format_dict(__lowerCAmelCase)) return info @staticmethod def a_ ( __lowerCAmelCase): """simple docstring""" return "\n".join([f"- {prop}: {val}" for prop, val in d.items()]) + "\n"
272
'''simple docstring''' from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig __lowercase = { '''susnato/ernie-m-base_pytorch''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json''', '''susnato/ernie-m-large_pytorch''': '''https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json''', } class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : List[str] = '''ernie_m''' UpperCAmelCase_ : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowerCAmelCase = 250002 , __lowerCAmelCase = 768 , __lowerCAmelCase = 12 , __lowerCAmelCase = 12 , __lowerCAmelCase = 3072 , __lowerCAmelCase = "gelu" , __lowerCAmelCase = 0.1 , __lowerCAmelCase = 0.1 , __lowerCAmelCase = 514 , __lowerCAmelCase = 0.02 , __lowerCAmelCase = 1 , __lowerCAmelCase = 1E-0_5 , __lowerCAmelCase=None , __lowerCAmelCase=False , __lowerCAmelCase=0.0 , **__lowerCAmelCase , ): """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = initializer_range lowerCAmelCase = layer_norm_eps lowerCAmelCase = classifier_dropout lowerCAmelCase = is_decoder lowerCAmelCase = act_dropout
272
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class a__( unittest.TestCase ): '''simple docstring''' def a_ ( self): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def a_ ( self): """simple docstring""" lowerCAmelCase = 1 lowerCAmelCase = 3 lowerCAmelCase = (32, 32) lowerCAmelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0)).to(__lowerCAmelCase) return image @property def a_ ( self): """simple docstring""" torch.manual_seed(0) lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) return model @property def a_ ( self): """simple docstring""" torch.manual_seed(0) lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def a_ ( self): """simple docstring""" torch.manual_seed(0) lowerCAmelCase = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(__lowerCAmelCase) @property def a_ ( self): """simple docstring""" def extract(*__lowerCAmelCase , **__lowerCAmelCase): class a__: '''simple docstring''' def __init__( self): """simple docstring""" lowerCAmelCase = torch.ones([0]) def a_ ( self , __lowerCAmelCase): """simple docstring""" self.pixel_values.to(__lowerCAmelCase) return self return Out() return extract def a_ ( self): """simple docstring""" lowerCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase = self.dummy_cond_unet lowerCAmelCase = PNDMScheduler(skip_prk_steps=__lowerCAmelCase) lowerCAmelCase = self.dummy_vae lowerCAmelCase = self.dummy_text_encoder lowerCAmelCase = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""") lowerCAmelCase = 77 lowerCAmelCase = self.dummy_image.to(__lowerCAmelCase) lowerCAmelCase = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk lowerCAmelCase = AltDiffusionImgaImgPipeline( unet=__lowerCAmelCase , scheduler=__lowerCAmelCase , vae=__lowerCAmelCase , text_encoder=__lowerCAmelCase , tokenizer=__lowerCAmelCase , safety_checker=__lowerCAmelCase , feature_extractor=self.dummy_extractor , ) lowerCAmelCase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=__lowerCAmelCase) lowerCAmelCase = alt_pipe.to(__lowerCAmelCase) alt_pipe.set_progress_bar_config(disable=__lowerCAmelCase) lowerCAmelCase = """A painting of a squirrel eating a burger""" lowerCAmelCase = torch.Generator(device=__lowerCAmelCase).manual_seed(0) lowerCAmelCase = alt_pipe( [prompt] , generator=__lowerCAmelCase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , image=__lowerCAmelCase , ) lowerCAmelCase = output.images lowerCAmelCase = torch.Generator(device=__lowerCAmelCase).manual_seed(0) lowerCAmelCase = alt_pipe( [prompt] , generator=__lowerCAmelCase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , image=__lowerCAmelCase , return_dict=__lowerCAmelCase , )[0] lowerCAmelCase = image[0, -3:, -3:, -1] lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase = np.array([0.4427, 0.3731, 0.4249, 0.4941, 0.4546, 0.4148, 0.4193, 0.4666, 0.4499]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 5E-3 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""") def a_ ( self): """simple docstring""" lowerCAmelCase = self.dummy_cond_unet lowerCAmelCase = PNDMScheduler(skip_prk_steps=__lowerCAmelCase) lowerCAmelCase = self.dummy_vae lowerCAmelCase = self.dummy_text_encoder lowerCAmelCase = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""") lowerCAmelCase = 77 lowerCAmelCase = self.dummy_image.to(__lowerCAmelCase) # put models in fp16 lowerCAmelCase = unet.half() lowerCAmelCase = vae.half() lowerCAmelCase = bert.half() # make sure here that pndm scheduler skips prk lowerCAmelCase = AltDiffusionImgaImgPipeline( unet=__lowerCAmelCase , scheduler=__lowerCAmelCase , vae=__lowerCAmelCase , text_encoder=__lowerCAmelCase , tokenizer=__lowerCAmelCase , safety_checker=__lowerCAmelCase , feature_extractor=self.dummy_extractor , ) lowerCAmelCase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=__lowerCAmelCase) lowerCAmelCase = alt_pipe.to(__lowerCAmelCase) alt_pipe.set_progress_bar_config(disable=__lowerCAmelCase) lowerCAmelCase = """A painting of a squirrel eating a burger""" lowerCAmelCase = torch.manual_seed(0) lowerCAmelCase = alt_pipe( [prompt] , generator=__lowerCAmelCase , num_inference_steps=2 , output_type="""np""" , image=__lowerCAmelCase , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""") def a_ ( self): """simple docstring""" lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""") # resize to resolution that is divisible by 8 but not 16 or 32 lowerCAmelCase = init_image.resize((760, 504)) lowerCAmelCase = """BAAI/AltDiffusion""" lowerCAmelCase = AltDiffusionImgaImgPipeline.from_pretrained( __lowerCAmelCase , safety_checker=__lowerCAmelCase , ) pipe.to(__lowerCAmelCase) pipe.set_progress_bar_config(disable=__lowerCAmelCase) pipe.enable_attention_slicing() lowerCAmelCase = """A fantasy landscape, trending on artstation""" lowerCAmelCase = torch.manual_seed(0) lowerCAmelCase = pipe( prompt=__lowerCAmelCase , image=__lowerCAmelCase , strength=0.75 , guidance_scale=7.5 , generator=__lowerCAmelCase , output_type="""np""" , ) lowerCAmelCase = output.images[0] lowerCAmelCase = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) lowerCAmelCase = np.array([0.9358, 0.9397, 0.9599, 0.9901, 1.0000, 1.0000, 0.9882, 1.0000, 1.0000]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 @slow @require_torch_gpu class a__( unittest.TestCase ): '''simple docstring''' def a_ ( self): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self): """simple docstring""" lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""") lowerCAmelCase = init_image.resize((768, 512)) lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy""") lowerCAmelCase = """BAAI/AltDiffusion""" lowerCAmelCase = AltDiffusionImgaImgPipeline.from_pretrained( __lowerCAmelCase , safety_checker=__lowerCAmelCase , ) pipe.to(__lowerCAmelCase) pipe.set_progress_bar_config(disable=__lowerCAmelCase) pipe.enable_attention_slicing() lowerCAmelCase = """A fantasy landscape, trending on artstation""" lowerCAmelCase = torch.manual_seed(0) lowerCAmelCase = pipe( prompt=__lowerCAmelCase , image=__lowerCAmelCase , strength=0.75 , guidance_scale=7.5 , generator=__lowerCAmelCase , output_type="""np""" , ) lowerCAmelCase = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image).max() < 1E-2
272
'''simple docstring''' import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __lowercase = logging.getLogger(__name__) class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : Any = '''sequence-classification''' def __init__( self , __lowerCAmelCase): """simple docstring""" if type(__lowerCAmelCase) == dict: lowerCAmelCase = Namespace(**__lowerCAmelCase) lowerCAmelCase = glue_output_modes[hparams.task] lowerCAmelCase = glue_tasks_num_labels[hparams.task] super().__init__(__lowerCAmelCase , __lowerCAmelCase , self.mode) def a_ ( self , **__lowerCAmelCase): """simple docstring""" return self.model(**__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowerCAmelCase = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None lowerCAmelCase = self(**__lowerCAmelCase) lowerCAmelCase = outputs[0] lowerCAmelCase = self.trainer.lr_schedulers[0]["""scheduler"""] lowerCAmelCase = {"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def a_ ( self): """simple docstring""" lowerCAmelCase = self.hparams lowerCAmelCase = processors[args.task]() lowerCAmelCase = processor.get_labels() for mode in ["train", "dev"]: lowerCAmelCase = self._feature_file(__lowerCAmelCase) if os.path.exists(__lowerCAmelCase) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , __lowerCAmelCase) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir) lowerCAmelCase = ( processor.get_dev_examples(args.data_dir) if mode == """dev""" else processor.get_train_examples(args.data_dir) ) lowerCAmelCase = convert_examples_to_features( __lowerCAmelCase , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("""Saving features into cached file %s""" , __lowerCAmelCase) torch.save(__lowerCAmelCase , __lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = False): """simple docstring""" lowerCAmelCase = """dev""" if mode == """test""" else mode lowerCAmelCase = self._feature_file(__lowerCAmelCase) logger.info("""Loading features from cached file %s""" , __lowerCAmelCase) lowerCAmelCase = torch.load(__lowerCAmelCase) lowerCAmelCase = torch.tensor([f.input_ids for f in features] , dtype=torch.long) lowerCAmelCase = torch.tensor([f.attention_mask for f in features] , dtype=torch.long) lowerCAmelCase = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long) if self.hparams.glue_output_mode == "classification": lowerCAmelCase = torch.tensor([f.label for f in features] , dtype=torch.long) elif self.hparams.glue_output_mode == "regression": lowerCAmelCase = torch.tensor([f.label for f in features] , dtype=torch.float) return DataLoader( TensorDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) , batch_size=__lowerCAmelCase , shuffle=__lowerCAmelCase , ) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowerCAmelCase = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None lowerCAmelCase = self(**__lowerCAmelCase) lowerCAmelCase , lowerCAmelCase = outputs[:2] lowerCAmelCase = logits.detach().cpu().numpy() lowerCAmelCase = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = torch.stack([x["""val_loss"""] for x in outputs]).mean().detach().cpu().item() lowerCAmelCase = np.concatenate([x["""pred"""] for x in outputs] , axis=0) if self.hparams.glue_output_mode == "classification": lowerCAmelCase = np.argmax(__lowerCAmelCase , axis=1) elif self.hparams.glue_output_mode == "regression": lowerCAmelCase = np.squeeze(__lowerCAmelCase) lowerCAmelCase = np.concatenate([x["""target"""] for x in outputs] , axis=0) lowerCAmelCase = [[] for _ in range(out_label_ids.shape[0])] lowerCAmelCase = [[] for _ in range(out_label_ids.shape[0])] lowerCAmelCase = {**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task , __lowerCAmelCase , __lowerCAmelCase)} lowerCAmelCase = dict(results.items()) lowerCAmelCase = results return ret, preds_list, out_label_list def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self._eval_end(__lowerCAmelCase) lowerCAmelCase = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self._eval_end(__lowerCAmelCase) lowerCAmelCase = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def a_ ( __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" BaseTransformer.add_model_specific_args(__lowerCAmelCase , __lowerCAmelCase) parser.add_argument( """--max_seq_length""" , default=128 , type=__lowerCAmelCase , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--task""" , default="""""" , type=__lowerCAmelCase , required=__lowerCAmelCase , help="""The GLUE task to run""" , ) parser.add_argument( """--gpus""" , default=0 , type=__lowerCAmelCase , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""") return parser def snake_case__ ( ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase = argparse.ArgumentParser() add_generic_args(_A , os.getcwd() ) lowerCAmelCase = GLUETransformer.add_model_specific_args(_A , os.getcwd() ) lowerCAmelCase = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: lowerCAmelCase = os.path.join( """./results""" , f"{args.task}_{time.strftime('%Y%m%d_%H%M%S' )}" , ) os.makedirs(args.output_dir ) lowerCAmelCase = GLUETransformer(_A ) lowerCAmelCase = generic_train(_A , _A ) # Optionally, predict on dev set and write to output_dir if args.do_predict: lowerCAmelCase = sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=_A ) ) lowerCAmelCase = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(_A ) if __name__ == "__main__": main()
272
1
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def snake_case__ ( _A: List[Any] ) -> List[Any]: '''simple docstring''' lowerCAmelCase = botoa.client("""iam""" ) lowerCAmelCase = { """Version""": """2012-10-17""", """Statement""": [ {"""Effect""": """Allow""", """Principal""": {"""Service""": """sagemaker.amazonaws.com"""}, """Action""": """sts:AssumeRole"""} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=_A , AssumeRolePolicyDocument=json.dumps(_A , indent=2 ) ) lowerCAmelCase = { """Version""": """2012-10-17""", """Statement""": [ { """Effect""": """Allow""", """Action""": [ """sagemaker:*""", """ecr:GetDownloadUrlForLayer""", """ecr:BatchGetImage""", """ecr:BatchCheckLayerAvailability""", """ecr:GetAuthorizationToken""", """cloudwatch:PutMetricData""", """cloudwatch:GetMetricData""", """cloudwatch:GetMetricStatistics""", """cloudwatch:ListMetrics""", """logs:CreateLogGroup""", """logs:CreateLogStream""", """logs:DescribeLogStreams""", """logs:PutLogEvents""", """logs:GetLogEvents""", """s3:CreateBucket""", """s3:ListBucket""", """s3:GetBucketLocation""", """s3:GetObject""", """s3:PutObject""", ], """Resource""": """*""", } ], } # attach policy to role iam_client.put_role_policy( RoleName=_A , PolicyName=f"{role_name}_policy_permission" , PolicyDocument=json.dumps(_A , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(f"role {role_name} already exists. Using existing one" ) def snake_case__ ( _A: Tuple ) -> Optional[int]: '''simple docstring''' lowerCAmelCase = botoa.client("""iam""" ) return iam_client.get_role(RoleName=_A )["Role"]["Arn"] def snake_case__ ( ) -> Tuple: '''simple docstring''' lowerCAmelCase = _ask_options( """How do you want to authorize?""" , ["""AWS Profile""", """Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) """] , _A , ) lowerCAmelCase = None if credentials_configuration == 0: lowerCAmelCase = _ask_field("""Enter your AWS Profile name: [default] """ , default="""default""" ) lowerCAmelCase = aws_profile else: print( """Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,""" """`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`""" ) lowerCAmelCase = _ask_field("""AWS Access Key ID: """ ) lowerCAmelCase = aws_access_key_id lowerCAmelCase = _ask_field("""AWS Secret Access Key: """ ) lowerCAmelCase = aws_secret_access_key lowerCAmelCase = _ask_field("""Enter your AWS Region: [us-east-1]""" , default="""us-east-1""" ) lowerCAmelCase = aws_region lowerCAmelCase = _ask_options( """Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?""" , ["""Provide IAM Role name""", """Create new IAM role using credentials"""] , _A , ) if role_management == 0: lowerCAmelCase = _ask_field("""Enter your IAM role name: """ ) else: lowerCAmelCase = """accelerate_sagemaker_execution_role""" print(f"Accelerate will create an iam role \"{iam_role_name}\" using the provided credentials" ) _create_iam_role_for_sagemaker(_A ) lowerCAmelCase = _ask_field( """Do you want to use custom Docker image? [yes/NO]: """ , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) lowerCAmelCase = None if is_custom_docker_image: lowerCAmelCase = _ask_field("""Enter your Docker image: """ , lambda _A : str(_A ).lower() ) lowerCAmelCase = _ask_field( """Do you want to provide SageMaker input channels with data locations? [yes/NO]: """ , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) lowerCAmelCase = None if is_sagemaker_inputs_enabled: lowerCAmelCase = _ask_field( """Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): """ , lambda _A : str(_A ).lower() , ) lowerCAmelCase = _ask_field( """Do you want to enable SageMaker metrics? [yes/NO]: """ , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) lowerCAmelCase = None if is_sagemaker_metrics_enabled: lowerCAmelCase = _ask_field( """Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): """ , lambda _A : str(_A ).lower() , ) lowerCAmelCase = _ask_options( """What is the distributed mode?""" , ["""No distributed training""", """Data parallelism"""] , _convert_sagemaker_distributed_mode , ) lowerCAmelCase = {} lowerCAmelCase = _ask_field( """Do you wish to optimize your script with torch dynamo?[yes/NO]:""" , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) if use_dynamo: lowerCAmelCase = """dynamo_""" lowerCAmelCase = _ask_options( """Which dynamo backend would you like to use?""" , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) lowerCAmelCase = _ask_field( """Do you want to customize the defaults sent to torch.compile? [yes/NO]: """ , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) if use_custom_options: lowerCAmelCase = _ask_options( """Which mode do you want to use?""" , _A , lambda _A : TORCH_DYNAMO_MODES[int(_A )] , default="""default""" , ) lowerCAmelCase = _ask_field( """Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: """ , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) lowerCAmelCase = _ask_field( """Do you want to enable dynamic shape tracing? [yes/NO]: """ , _convert_yes_no_to_bool , default=_A , error_message="""Please enter yes or no.""" , ) lowerCAmelCase = """Which EC2 instance type you want to use for your training?""" if distributed_type != SageMakerDistributedType.NO: lowerCAmelCase = _ask_options( _A , _A , lambda _A : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(_A )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" lowerCAmelCase = _ask_field(_A , lambda _A : str(_A ).lower() , default="""ml.p3.2xlarge""" ) lowerCAmelCase = 1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): lowerCAmelCase = _ask_field( """How many machines do you want use? [1]: """ , _A , default=1 , ) lowerCAmelCase = _ask_options( """Do you wish to use FP16 or BF16 (mixed precision)?""" , ["""no""", """fp16""", """bf16""", """fp8"""] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( """Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.""" ) return SageMakerConfig( image_uri=_A , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=_A , use_cpu=_A , dynamo_config=_A , eca_instance_type=_A , profile=_A , region=_A , iam_role_name=_A , mixed_precision=_A , num_machines=_A , sagemaker_inputs_file=_A , sagemaker_metrics_file=_A , )
272
'''simple docstring''' import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor __lowercase = logging.get_logger(__name__) class a__( lowerCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" warnings.warn( """The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use DeformableDetrImageProcessor instead.""" , __lowerCAmelCase , ) super().__init__(*__lowerCAmelCase , **__lowerCAmelCase)
272
1
'''simple docstring''' def snake_case__ ( _A: int ) -> None: '''simple docstring''' lowerCAmelCase = generate_pascal_triangle(_A ) for row_idx in range(_A ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=""" """ ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=""" """ ) else: print(triangle[row_idx][col_idx] , end="""""" ) print() def snake_case__ ( _A: int ) -> list[list[int]]: '''simple docstring''' if not isinstance(_A , _A ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) lowerCAmelCase = [] for current_row_idx in range(_A ): lowerCAmelCase = populate_current_row(_A , _A ) triangle.append(_A ) return triangle def snake_case__ ( _A: list[list[int]] , _A: int ) -> list[int]: '''simple docstring''' lowerCAmelCase = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 lowerCAmelCase , lowerCAmelCase = 1, 1 for current_col_idx in range(1 , _A ): calculate_current_element( _A , _A , _A , _A ) return current_row def snake_case__ ( _A: list[list[int]] , _A: list[int] , _A: int , _A: int , ) -> None: '''simple docstring''' lowerCAmelCase = triangle[current_row_idx - 1][current_col_idx - 1] lowerCAmelCase = triangle[current_row_idx - 1][current_col_idx] lowerCAmelCase = above_to_left_elt + above_to_right_elt def snake_case__ ( _A: int ) -> list[list[int]]: '''simple docstring''' if not isinstance(_A , _A ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) lowerCAmelCase = [[1]] for row_index in range(1 , _A ): lowerCAmelCase = [0] + result[-1] + [0] lowerCAmelCase = row_index + 1 # Calculate the number of distinct elements in a row lowerCAmelCase = sum(divmod(_A , 2 ) ) lowerCAmelCase = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] lowerCAmelCase = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() lowerCAmelCase = row_first_half + row_second_half result.append(_A ) return result def snake_case__ ( ) -> None: '''simple docstring''' from collections.abc import Callable from timeit import timeit def benchmark_a_function(_A: Callable , _A: int ) -> None: lowerCAmelCase = f"{func.__name__}({value})" lowerCAmelCase = timeit(f"__main__.{call}" , setup="""import __main__""" ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(f"{call:38} -- {timing:.4f} seconds" ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(_A , _A ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
272
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowercase = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
272
1
'''simple docstring''' from functools import lru_cache @lru_cache def snake_case__ ( _A: int ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
272
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class a__( unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Dict = ViTImageProcessor if is_vision_available() else None @property def a_ ( self): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self): """simple docstring""" lowerCAmelCase = (3, 32, 128) lowerCAmelCase = tempfile.mkdtemp() # fmt: off lowerCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on lowerCAmelCase = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase)))) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""]) with open(self.vocab_file , """w""" , encoding="""utf-8""") as fp: fp.write(json.dumps(__lowerCAmelCase) + """\n""") lowerCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 32, """width""": 128}, } lowerCAmelCase = os.path.join(self.tmpdirname , __lowerCAmelCase) with open(self.image_processor_file , """w""" , encoding="""utf-8""") as fp: json.dump(__lowerCAmelCase , __lowerCAmelCase) def a_ ( self , **__lowerCAmelCase): """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCAmelCase) def a_ ( self , **__lowerCAmelCase): """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCAmelCase) def a_ ( self): """simple docstring""" shutil.rmtree(self.tmpdirname) def a_ ( self): """simple docstring""" lowerCAmelCase = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta) lowerCAmelCase = Image.fromarray(np.moveaxis(__lowerCAmelCase , 0 , -1)) return image_input def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_image_processor() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) processor.save_pretrained(self.tmpdirname) lowerCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCAmelCase) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab()) self.assertIsInstance(processor.char_tokenizer , __lowerCAmelCase) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string()) self.assertIsInstance(processor.image_processor , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_image_processor() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) processor.save_pretrained(self.tmpdirname) lowerCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""") lowerCAmelCase = self.get_image_processor(do_normalize=__lowerCAmelCase , padding_value=1.0) lowerCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCAmelCase , padding_value=1.0) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.char_tokenizer , __lowerCAmelCase) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = self.prepare_image_inputs() lowerCAmelCase = image_processor(__lowerCAmelCase , return_tensors="""np""") lowerCAmelCase = processor(images=__lowerCAmelCase , return_tensors="""np""") for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = """test""" lowerCAmelCase = processor(text=__lowerCAmelCase) lowerCAmelCase = tokenizer(__lowerCAmelCase) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = """test""" lowerCAmelCase = self.prepare_image_inputs() lowerCAmelCase = processor(text=__lowerCAmelCase , images=__lowerCAmelCase) self.assertListEqual(list(inputs.keys()) , ["""pixel_values""", """labels"""]) # test if it raises when no input is passed with pytest.raises(__lowerCAmelCase): processor() def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase = processor.char_decode(__lowerCAmelCase) lowerCAmelCase = tokenizer.batch_decode(__lowerCAmelCase) lowerCAmelCase = [seq.replace(""" """ , """""") for seq in decoded_tok] self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = None lowerCAmelCase = self.prepare_image_inputs() lowerCAmelCase = processor(text=__lowerCAmelCase , images=__lowerCAmelCase) self.assertListEqual(list(inputs.keys()) , processor.model_input_names) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = torch.randn(1 , 27 , 38) lowerCAmelCase = torch.randn(1 , 27 , 50257) lowerCAmelCase = torch.randn(1 , 27 , 30522) lowerCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input]) self.assertListEqual(list(results.keys()) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""])
272
1
'''simple docstring''' import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__: '''simple docstring''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=3 , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=224 , __lowerCAmelCase=1000 , __lowerCAmelCase=[3, 3, 6, 4] , __lowerCAmelCase=[48, 56, 112, 220] , ): """simple docstring""" lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = num_channels lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = num_labels lowerCAmelCase = image_size lowerCAmelCase = layer_depths lowerCAmelCase = embed_dims def a_ ( self): """simple docstring""" lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.num_labels) lowerCAmelCase = self.get_config() return config, pixel_values, labels def a_ ( self): """simple docstring""" return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="""gelu""" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=__lowerCAmelCase , layer_scale_init_value=1E-5 , ) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = SwiftFormerModel(config=__lowerCAmelCase) model.to(__lowerCAmelCase) model.eval() lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.num_labels lowerCAmelCase = SwiftFormerForImageClassification(__lowerCAmelCase) model.to(__lowerCAmelCase) model.eval() lowerCAmelCase = model(__lowerCAmelCase , labels=__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) lowerCAmelCase = SwiftFormerForImageClassification(__lowerCAmelCase) model.to(__lowerCAmelCase) model.eval() lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def a_ ( self): """simple docstring""" ((lowerCAmelCase) , (lowerCAmelCase) , (lowerCAmelCase)) = self.prepare_config_and_inputs() lowerCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a__( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Dict = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () UpperCAmelCase_ : List[str] = ( {'''feature-extraction''': SwiftFormerModel, '''image-classification''': SwiftFormerForImageClassification} if is_torch_available() else {} ) UpperCAmelCase_ : Union[str, Any] = False UpperCAmelCase_ : Optional[int] = False UpperCAmelCase_ : Optional[Any] = False UpperCAmelCase_ : Any = False UpperCAmelCase_ : Optional[Any] = False def a_ ( self): """simple docstring""" lowerCAmelCase = SwiftFormerModelTester(self) lowerCAmelCase = ConfigTester( self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def a_ ( self): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""SwiftFormer does not use inputs_embeds""") def a_ ( self): """simple docstring""" pass def a_ ( self): """simple docstring""" lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(__lowerCAmelCase) lowerCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear)) def a_ ( self): """simple docstring""" lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(__lowerCAmelCase) lowerCAmelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCAmelCase) @slow def a_ ( self): """simple docstring""" for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = SwiftFormerModel.from_pretrained(__lowerCAmelCase) self.assertIsNotNone(__lowerCAmelCase) @unittest.skip(reason="""SwiftFormer does not output attentions""") def a_ ( self): """simple docstring""" pass def a_ ( self): """simple docstring""" def check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): lowerCAmelCase = model_class(__lowerCAmelCase) model.to(__lowerCAmelCase) model.eval() with torch.no_grad(): lowerCAmelCase = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase)) lowerCAmelCase = outputs.hidden_states lowerCAmelCase = 8 self.assertEqual(len(__lowerCAmelCase) , __lowerCAmelCase) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(__lowerCAmelCase)): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ]) , ) lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) def a_ ( self): """simple docstring""" def _config_zero_init(__lowerCAmelCase): lowerCAmelCase = copy.deepcopy(__lowerCAmelCase) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(__lowerCAmelCase , __lowerCAmelCase , 1E-1_0) if isinstance(getattr(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) , __lowerCAmelCase): lowerCAmelCase = _config_zero_init(getattr(__lowerCAmelCase , __lowerCAmelCase)) setattr(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) return configs_no_init lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = _config_zero_init(__lowerCAmelCase) for model_class in self.all_model_classes: lowerCAmelCase = model_class(config=__lowerCAmelCase) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=f"Parameter {name} of model {model_class} seems not properly initialized" , ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""") def a_ ( self): """simple docstring""" pass def snake_case__ ( ) -> Any: '''simple docstring''' lowerCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class a__( unittest.TestCase ): '''simple docstring''' @cached_property def a_ ( self): """simple docstring""" return ViTImageProcessor.from_pretrained("""MBZUAI/swiftformer-xs""") if is_vision_available() else None @slow def a_ ( self): """simple docstring""" lowerCAmelCase = SwiftFormerForImageClassification.from_pretrained("""MBZUAI/swiftformer-xs""").to(__lowerCAmelCase) lowerCAmelCase = self.default_image_processor lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=__lowerCAmelCase , return_tensors="""pt""").to(__lowerCAmelCase) # forward pass with torch.no_grad(): lowerCAmelCase = model(**__lowerCAmelCase) # verify the logits lowerCAmelCase = torch.Size((1, 1000)) self.assertEqual(outputs.logits.shape , __lowerCAmelCase) lowerCAmelCase = torch.tensor([[-2.1_7_0_3E0_0, 2.1_1_0_7E0_0, -2.0_8_1_1E0_0]]).to(__lowerCAmelCase) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1E-4))
272
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __lowercase = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class a__( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Tuple = XLMRobertaTokenizer UpperCAmelCase_ : int = XLMRobertaTokenizerFast UpperCAmelCase_ : List[str] = True UpperCAmelCase_ : Optional[int] = True def a_ ( self): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase = XLMRobertaTokenizer(__lowerCAmelCase , keep_accents=__lowerCAmelCase) tokenizer.save_pretrained(self.tmpdirname) def a_ ( self): """simple docstring""" lowerCAmelCase = """<pad>""" lowerCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCAmelCase) , __lowerCAmelCase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCAmelCase) , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , """<s>""") self.assertEqual(vocab_keys[1] , """<pad>""") self.assertEqual(vocab_keys[-1] , """<mask>""") self.assertEqual(len(__lowerCAmelCase) , 1002) def a_ ( self): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1002) def a_ ( self): """simple docstring""" lowerCAmelCase = XLMRobertaTokenizer(__lowerCAmelCase , keep_accents=__lowerCAmelCase) lowerCAmelCase = tokenizer.tokenize("""This is a test""") self.assertListEqual(__lowerCAmelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""]) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCAmelCase) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""") self.assertListEqual( __lowerCAmelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) lowerCAmelCase = tokenizer.convert_tokens_to_ids(__lowerCAmelCase) self.assertListEqual( __lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase = tokenizer.convert_ids_to_tokens(__lowerCAmelCase) self.assertListEqual( __lowerCAmelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def a_ ( self): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})"): lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = self.tokenizer_class.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(__lowerCAmelCase) lowerCAmelCase = tokenizer_p.save_pretrained(__lowerCAmelCase) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files)) lowerCAmelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f) self.assertSequenceEqual(__lowerCAmelCase , __lowerCAmelCase) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(__lowerCAmelCase) lowerCAmelCase = tokenizer_p.from_pretrained(__lowerCAmelCase) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCAmelCase , __lowerCAmelCase)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowerCAmelCase) # Save tokenizer rust, legacy_format=True lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(__lowerCAmelCase , legacy_format=__lowerCAmelCase) lowerCAmelCase = tokenizer_p.save_pretrained(__lowerCAmelCase) # Checks it save with the same files self.assertSequenceEqual(__lowerCAmelCase , __lowerCAmelCase) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(__lowerCAmelCase) lowerCAmelCase = tokenizer_p.from_pretrained(__lowerCAmelCase) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCAmelCase , __lowerCAmelCase)) shutil.rmtree(__lowerCAmelCase) # Save tokenizer rust, legacy_format=False lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(__lowerCAmelCase , legacy_format=__lowerCAmelCase) lowerCAmelCase = tokenizer_p.save_pretrained(__lowerCAmelCase) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(__lowerCAmelCase) lowerCAmelCase = tokenizer_p.from_pretrained(__lowerCAmelCase) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCAmelCase , __lowerCAmelCase)) shutil.rmtree(__lowerCAmelCase) @cached_property def a_ ( self): """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""") def a_ ( self): """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__lowerCAmelCase , f.name) lowerCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=__lowerCAmelCase) lowerCAmelCase = pickle.dumps(__lowerCAmelCase) pickle.loads(__lowerCAmelCase) def a_ ( self): """simple docstring""" if not self.test_rust_tokenizer: return lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_rust_tokenizer() lowerCAmelCase = """I was born in 92000, and this is falsé.""" lowerCAmelCase = tokenizer.tokenize(__lowerCAmelCase) lowerCAmelCase = rust_tokenizer.tokenize(__lowerCAmelCase) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase) lowerCAmelCase = rust_tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = self.get_rust_tokenizer() lowerCAmelCase = tokenizer.encode(__lowerCAmelCase) lowerCAmelCase = rust_tokenizer.encode(__lowerCAmelCase) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase) @slow def a_ ( self): """simple docstring""" lowerCAmelCase = """Hello World!""" lowerCAmelCase = [0, 35378, 6661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__lowerCAmelCase , self.big_tokenizer.encode(__lowerCAmelCase)) @slow def a_ ( self): """simple docstring""" lowerCAmelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) lowerCAmelCase = [ 0, 3293, 83, 10, 4552, 4989, 7986, 678, 10, 5915, 111, 179459, 124850, 4, 6044, 237, 12, 6, 5, 6, 4, 6780, 705, 15, 1388, 44, 378, 10114, 711, 152, 20, 6, 5, 22376, 642, 1221, 15190, 34153, 450, 5608, 959, 1119, 57702, 136, 186, 47, 1098, 29367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6044, 237, 6284, 50901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__lowerCAmelCase , self.big_tokenizer.encode(__lowerCAmelCase)) @slow def a_ ( self): """simple docstring""" lowerCAmelCase = {"""input_ids""": [[0, 11062, 82772, 7, 15, 82772, 538, 51529, 237, 17198, 1290, 206, 9, 215175, 1314, 136, 17198, 1290, 206, 9, 56359, 42, 122009, 9, 16466, 16, 87344, 4537, 9, 4717, 78381, 6, 159958, 7, 15, 24480, 618, 4, 527, 22693, 5428, 4, 2777, 24480, 9874, 4, 43523, 594, 4, 803, 18392, 33189, 18, 4, 43523, 24447, 12399, 100, 24955, 83658, 9626, 144057, 15, 839, 22335, 16, 136, 24955, 83658, 83479, 15, 39102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 122009, 115774, 23, 805, 1328, 46876, 7, 136, 53894, 1940, 42227, 41159, 17721, 823, 425, 4, 27512, 98722, 206, 136, 5531, 4970, 919, 17336, 5, 2], [0, 20080, 618, 83, 82775, 47, 479, 9, 1517, 73, 53894, 333, 80581, 110117, 18811, 5256, 1295, 51, 152526, 297, 7986, 390, 124416, 538, 35431, 214, 98, 15044, 25737, 136, 7108, 43701, 23, 756, 135355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63773, 119455, 6, 147797, 88203, 7, 645, 70, 21, 3285, 10269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCAmelCase , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
272
1
'''simple docstring''' def snake_case__ ( _A: int ) -> int: '''simple docstring''' if not isinstance(_A , _A ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
272
'''simple docstring''' def snake_case__ ( _A: int , _A: int ) -> int: '''simple docstring''' while a != 0: lowerCAmelCase , lowerCAmelCase = b % a, a return b def snake_case__ ( _A: int , _A: int ) -> int: '''simple docstring''' if gcd(_A , _A ) != 1: lowerCAmelCase = f"mod inverse of {a!r} and {m!r} does not exist" raise ValueError(_A ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 1, 0, a lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 0, 1, m while va != 0: lowerCAmelCase = ua // va lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
272
1
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class a__( lowerCAmelCase__ ): '''simple docstring''' @staticmethod @abstractmethod def a_ ( __lowerCAmelCase): """simple docstring""" raise NotImplementedError() @abstractmethod def a_ ( self): """simple docstring""" raise NotImplementedError()
272
'''simple docstring''' import math import flax.linen as nn import jax.numpy as jnp def snake_case__ ( _A: jnp.ndarray , _A: int , _A: float = 1 , _A: float = 1 , _A: float = 1.0e4 , _A: bool = False , _A: float = 1.0 , ) -> jnp.ndarray: '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"Embedding dimension {embedding_dim} should be even" lowerCAmelCase = float(embedding_dim // 2 ) lowerCAmelCase = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowerCAmelCase = min_timescale * jnp.exp(jnp.arange(_A , dtype=jnp.floataa ) * -log_timescale_increment ) lowerCAmelCase = jnp.expand_dims(_A , 1 ) * jnp.expand_dims(_A , 0 ) # scale embeddings lowerCAmelCase = scale * emb if flip_sin_to_cos: lowerCAmelCase = jnp.concatenate([jnp.cos(_A ), jnp.sin(_A )] , axis=1 ) else: lowerCAmelCase = jnp.concatenate([jnp.sin(_A ), jnp.cos(_A )] , axis=1 ) lowerCAmelCase = jnp.reshape(_A , [jnp.shape(_A )[0], embedding_dim] ) return signal class a__( nn.Module ): '''simple docstring''' UpperCAmelCase_ : int = 3_2 UpperCAmelCase_ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = nn.Dense(self.time_embed_dim , dtype=self.dtype , name="""linear_1""")(__lowerCAmelCase) lowerCAmelCase = nn.silu(__lowerCAmelCase) lowerCAmelCase = nn.Dense(self.time_embed_dim , dtype=self.dtype , name="""linear_2""")(__lowerCAmelCase) return temb class a__( nn.Module ): '''simple docstring''' UpperCAmelCase_ : int = 3_2 UpperCAmelCase_ : bool = False UpperCAmelCase_ : float = 1 @nn.compact def __call__( self , __lowerCAmelCase): """simple docstring""" return get_sinusoidal_embeddings( __lowerCAmelCase , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift)
272
1
'''simple docstring''' def snake_case__ ( _A: str ) -> bool: '''simple docstring''' lowerCAmelCase = 0 for ch in input_str: lowerCAmelCase = ord(_A ) lowerCAmelCase = pow(2 , _A ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
272
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowercase = { '''configuration_nezha''': ['''NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''NezhaConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''NezhaForNextSentencePrediction''', '''NezhaForMaskedLM''', '''NezhaForPreTraining''', '''NezhaForMultipleChoice''', '''NezhaForQuestionAnswering''', '''NezhaForSequenceClassification''', '''NezhaForTokenClassification''', '''NezhaModel''', '''NezhaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
272
1
'''simple docstring''' from sklearn.metrics import matthews_corrcoef import datasets __lowercase = ''' Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] ''' __lowercase = ''' Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results[\'matthews_correlation\'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results[\'matthews_correlation\'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results[\'matthews_correlation\'], 2)) -0.25 ''' __lowercase = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__( datasets.Metric ): '''simple docstring''' def a_ ( self): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32"""), """references""": datasets.Value("""int32"""), }) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None): """simple docstring""" return { "matthews_correlation": float(matthews_corrcoef(__lowerCAmelCase , __lowerCAmelCase , sample_weight=__lowerCAmelCase)), }
272
'''simple docstring''' from math import sqrt def snake_case__ ( _A: int = 1000000 ) -> int: '''simple docstring''' lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_A , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'{solution() = }')
272
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowercase = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
272
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowercase = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
272
1
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class a__( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = IFInpaintingPipeline UpperCAmelCase_ : Dict = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} UpperCAmelCase_ : Tuple = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS UpperCAmelCase_ : Dict = PipelineTesterMixin.required_optional_params - {'''latents'''} def a_ ( self): """simple docstring""" return self._get_dummy_components() def a_ ( self , __lowerCAmelCase , __lowerCAmelCase=0): """simple docstring""" if str(__lowerCAmelCase).startswith("""mps"""): lowerCAmelCase = torch.manual_seed(__lowerCAmelCase) else: lowerCAmelCase = torch.Generator(device=__lowerCAmelCase).manual_seed(__lowerCAmelCase) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowerCAmelCase)).to(__lowerCAmelCase) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowerCAmelCase)).to(__lowerCAmelCase) lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def a_ ( self): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3) def a_ ( self): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""") def a_ ( self): """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1) def a_ ( self): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2) def a_ ( self): """simple docstring""" self._test_save_load_local() def a_ ( self): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
272
'''simple docstring''' import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class a__( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase=7 , __lowerCAmelCase=3 , __lowerCAmelCase=18 , __lowerCAmelCase=30 , __lowerCAmelCase=400 , __lowerCAmelCase=True , __lowerCAmelCase=None , __lowerCAmelCase=True , __lowerCAmelCase=[0.5, 0.5, 0.5] , __lowerCAmelCase=[0.5, 0.5, 0.5] , ): """simple docstring""" lowerCAmelCase = size if size is not None else {"""height""": 18, """width""": 18} lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = num_channels lowerCAmelCase = image_size lowerCAmelCase = min_resolution lowerCAmelCase = max_resolution lowerCAmelCase = do_resize lowerCAmelCase = size lowerCAmelCase = do_normalize lowerCAmelCase = image_mean lowerCAmelCase = image_std def a_ ( self): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class a__( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Tuple = DPTImageProcessor if is_vision_available() else None def a_ ( self): """simple docstring""" lowerCAmelCase = DPTImageProcessingTester(self) @property def a_ ( self): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self): """simple docstring""" lowerCAmelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__lowerCAmelCase , """image_mean""")) self.assertTrue(hasattr(__lowerCAmelCase , """image_std""")) self.assertTrue(hasattr(__lowerCAmelCase , """do_normalize""")) self.assertTrue(hasattr(__lowerCAmelCase , """do_resize""")) self.assertTrue(hasattr(__lowerCAmelCase , """size""")) def a_ ( self): """simple docstring""" lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18}) lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42}) def a_ ( self): """simple docstring""" lowerCAmelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , Image.Image) # Test not batched input lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCAmelCase = image_processing(__lowerCAmelCase , return_tensors="""pt""").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def a_ ( self): """simple docstring""" lowerCAmelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase , numpify=__lowerCAmelCase) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , np.ndarray) # Test not batched input lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCAmelCase = image_processing(__lowerCAmelCase , return_tensors="""pt""").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def a_ ( self): """simple docstring""" lowerCAmelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase , torchify=__lowerCAmelCase) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , torch.Tensor) # Test not batched input lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCAmelCase = image_processing(__lowerCAmelCase , return_tensors="""pt""").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
272
1
'''simple docstring''' from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class a__( lowerCAmelCase__ ): '''simple docstring''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = False , __lowerCAmelCase = False , __lowerCAmelCase = None , **__lowerCAmelCase , ): """simple docstring""" super().__init__( __lowerCAmelCase , split=__lowerCAmelCase , features=__lowerCAmelCase , cache_dir=__lowerCAmelCase , keep_in_memory=__lowerCAmelCase , streaming=__lowerCAmelCase , num_proc=__lowerCAmelCase , **__lowerCAmelCase , ) lowerCAmelCase = path_or_paths if isinstance(__lowerCAmelCase , __lowerCAmelCase) else {self.split: path_or_paths} lowerCAmelCase = Text( cache_dir=__lowerCAmelCase , data_files=__lowerCAmelCase , features=__lowerCAmelCase , **__lowerCAmelCase , ) def a_ ( self): """simple docstring""" if self.streaming: lowerCAmelCase = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None self.builder.download_and_prepare( download_config=__lowerCAmelCase , download_mode=__lowerCAmelCase , verification_mode=__lowerCAmelCase , base_path=__lowerCAmelCase , num_proc=self.num_proc , ) lowerCAmelCase = self.builder.as_dataset( split=self.split , verification_mode=__lowerCAmelCase , in_memory=self.keep_in_memory) return dataset
272
'''simple docstring''' from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def snake_case__ ( _A: Union[str, Any] , _A: Tuple , _A: Any=1e-12 ) -> str: '''simple docstring''' lowerCAmelCase = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(_A , axis=1 ) , a_min=_A ) ).T lowerCAmelCase = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(_A , axis=1 ) , a_min=_A ) ).T return jnp.matmul(_A , norm_emb_a.T ) class a__( nn.Module ): '''simple docstring''' UpperCAmelCase_ : CLIPConfig UpperCAmelCase_ : jnp.dtype = jnp.floataa def a_ ( self): """simple docstring""" lowerCAmelCase = FlaxCLIPVisionModule(self.config.vision_config) lowerCAmelCase = nn.Dense(self.config.projection_dim , use_bias=__lowerCAmelCase , dtype=self.dtype) lowerCAmelCase = self.param("""concept_embeds""" , jax.nn.initializers.ones , (17, self.config.projection_dim)) lowerCAmelCase = self.param( """special_care_embeds""" , jax.nn.initializers.ones , (3, self.config.projection_dim)) lowerCAmelCase = self.param("""concept_embeds_weights""" , jax.nn.initializers.ones , (17,)) lowerCAmelCase = self.param("""special_care_embeds_weights""" , jax.nn.initializers.ones , (3,)) def __call__( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.vision_model(__lowerCAmelCase)[1] lowerCAmelCase = self.visual_projection(__lowerCAmelCase) lowerCAmelCase = jax_cosine_distance(__lowerCAmelCase , self.special_care_embeds) lowerCAmelCase = jax_cosine_distance(__lowerCAmelCase , self.concept_embeds) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs lowerCAmelCase = 0.0 lowerCAmelCase = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment lowerCAmelCase = jnp.round(__lowerCAmelCase , 3) lowerCAmelCase = jnp.any(special_scores > 0 , axis=1 , keepdims=__lowerCAmelCase) # Use a lower threshold if an image has any special care concept lowerCAmelCase = is_special_care * 0.01 lowerCAmelCase = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment lowerCAmelCase = jnp.round(__lowerCAmelCase , 3) lowerCAmelCase = jnp.any(concept_scores > 0 , axis=1) return has_nsfw_concepts class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : int = CLIPConfig UpperCAmelCase_ : Any = '''clip_input''' UpperCAmelCase_ : List[str] = FlaxStableDiffusionSafetyCheckerModule def __init__( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = 0 , __lowerCAmelCase = jnp.floataa , __lowerCAmelCase = True , **__lowerCAmelCase , ): """simple docstring""" if input_shape is None: lowerCAmelCase = (1, 224, 224, 3) lowerCAmelCase = self.module_class(config=__lowerCAmelCase , dtype=__lowerCAmelCase , **__lowerCAmelCase) super().__init__(__lowerCAmelCase , __lowerCAmelCase , input_shape=__lowerCAmelCase , seed=__lowerCAmelCase , dtype=__lowerCAmelCase , _do_init=_do_init) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = None): """simple docstring""" lowerCAmelCase = jax.random.normal(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase , lowerCAmelCase = jax.random.split(__lowerCAmelCase) lowerCAmelCase = {"""params""": params_rng, """dropout""": dropout_rng} lowerCAmelCase = self.module.init(__lowerCAmelCase , __lowerCAmelCase)["""params"""] return random_params def __call__( self , __lowerCAmelCase , __lowerCAmelCase = None , ): """simple docstring""" lowerCAmelCase = jnp.transpose(__lowerCAmelCase , (0, 2, 3, 1)) return self.module.apply( {"""params""": params or self.params} , jnp.array(__lowerCAmelCase , dtype=jnp.floataa) , rngs={} , )
272
1
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def snake_case__ ( _A: Tuple , _A: Dict=0.999 , _A: str="cosine" , ) -> int: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(_A: Union[str, Any] ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_A: List[str] ): return math.exp(t * -12.0 ) else: raise ValueError(f"Unsupported alpha_tranform_type: {alpha_transform_type}" ) lowerCAmelCase = [] for i in range(_A ): lowerCAmelCase = i / num_diffusion_timesteps lowerCAmelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(_A ) / alpha_bar_fn(_A ) , _A ) ) return torch.tensor(_A , dtype=torch.floataa ) class a__( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : Any = [e.name for e in KarrasDiffusionSchedulers] UpperCAmelCase_ : Optional[Any] = 2 @register_to_config def __init__( self , __lowerCAmelCase = 1000 , __lowerCAmelCase = 0.00085 , __lowerCAmelCase = 0.012 , __lowerCAmelCase = "linear" , __lowerCAmelCase = None , __lowerCAmelCase = "epsilon" , __lowerCAmelCase = "linspace" , __lowerCAmelCase = 0 , ): """simple docstring""" if trained_betas is not None: lowerCAmelCase = torch.tensor(__lowerCAmelCase , dtype=torch.floataa) elif beta_schedule == "linear": lowerCAmelCase = torch.linspace(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , dtype=torch.floataa) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. lowerCAmelCase = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , __lowerCAmelCase , dtype=torch.floataa) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule lowerCAmelCase = betas_for_alpha_bar(__lowerCAmelCase) else: raise NotImplementedError(f"{beta_schedule} does is not implemented for {self.__class__}") lowerCAmelCase = 1.0 - self.betas lowerCAmelCase = torch.cumprod(self.alphas , dim=0) # set all values self.set_timesteps(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase=None): """simple docstring""" if schedule_timesteps is None: lowerCAmelCase = self.timesteps lowerCAmelCase = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter) == 0: lowerCAmelCase = 1 if len(__lowerCAmelCase) > 1 else 0 else: lowerCAmelCase = timestep.cpu().item() if torch.is_tensor(__lowerCAmelCase) else timestep lowerCAmelCase = self._index_counter[timestep_int] return indices[pos].item() @property def a_ ( self): """simple docstring""" if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , ): """simple docstring""" lowerCAmelCase = self.index_for_timestep(__lowerCAmelCase) if self.state_in_first_order: lowerCAmelCase = self.sigmas[step_index] else: lowerCAmelCase = self.sigmas_interpol[step_index] lowerCAmelCase = sample / ((sigma**2 + 1) ** 0.5) return sample def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = None , ): """simple docstring""" lowerCAmelCase = num_inference_steps lowerCAmelCase = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": lowerCAmelCase = np.linspace(0 , num_train_timesteps - 1 , __lowerCAmelCase , dtype=__lowerCAmelCase)[::-1].copy() elif self.config.timestep_spacing == "leading": lowerCAmelCase = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 lowerCAmelCase = (np.arange(0 , __lowerCAmelCase) * step_ratio).round()[::-1].copy().astype(__lowerCAmelCase) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": lowerCAmelCase = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 lowerCAmelCase = (np.arange(__lowerCAmelCase , 0 , -step_ratio)).round().copy().astype(__lowerCAmelCase) timesteps -= 1 else: raise ValueError( f"{self.config.timestep_spacing} is not supported. Please make sure to choose one of 'linspace', 'leading' or 'trailing'.") lowerCAmelCase = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5) lowerCAmelCase = torch.from_numpy(np.log(__lowerCAmelCase)).to(__lowerCAmelCase) lowerCAmelCase = np.interp(__lowerCAmelCase , np.arange(0 , len(__lowerCAmelCase)) , __lowerCAmelCase) lowerCAmelCase = np.concatenate([sigmas, [0.0]]).astype(np.floataa) lowerCAmelCase = torch.from_numpy(__lowerCAmelCase).to(device=__lowerCAmelCase) # interpolate sigmas lowerCAmelCase = sigmas.log().lerp(sigmas.roll(1).log() , 0.5).exp() lowerCAmelCase = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2), sigmas[-1:]]) lowerCAmelCase = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2), sigmas_interpol[-1:]]) if str(__lowerCAmelCase).startswith("""mps"""): # mps does not support float64 lowerCAmelCase = torch.from_numpy(__lowerCAmelCase).to(__lowerCAmelCase , dtype=torch.floataa) else: lowerCAmelCase = torch.from_numpy(__lowerCAmelCase).to(__lowerCAmelCase) # interpolate timesteps lowerCAmelCase = self.sigma_to_t(__lowerCAmelCase).to(__lowerCAmelCase , dtype=timesteps.dtype) lowerCAmelCase = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1).flatten() lowerCAmelCase = torch.cat([timesteps[:1], interleaved_timesteps]) lowerCAmelCase = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter lowerCAmelCase = defaultdict(__lowerCAmelCase) def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = sigma.log() # get distribution lowerCAmelCase = log_sigma - self.log_sigmas[:, None] # get sigmas range lowerCAmelCase = dists.ge(0).cumsum(dim=0).argmax(dim=0).clamp(max=self.log_sigmas.shape[0] - 2) lowerCAmelCase = low_idx + 1 lowerCAmelCase = self.log_sigmas[low_idx] lowerCAmelCase = self.log_sigmas[high_idx] # interpolate sigmas lowerCAmelCase = (low - log_sigma) / (low - high) lowerCAmelCase = w.clamp(0 , 1) # transform interpolation to time range lowerCAmelCase = (1 - w) * low_idx + w * high_idx lowerCAmelCase = t.view(sigma.shape) return t @property def a_ ( self): """simple docstring""" return self.sample is None def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = True , ): """simple docstring""" lowerCAmelCase = self.index_for_timestep(__lowerCAmelCase) # advance index counter by 1 lowerCAmelCase = timestep.cpu().item() if torch.is_tensor(__lowerCAmelCase) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: lowerCAmelCase = self.sigmas[step_index] lowerCAmelCase = self.sigmas_interpol[step_index + 1] lowerCAmelCase = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method lowerCAmelCase = self.sigmas[step_index - 1] lowerCAmelCase = self.sigmas_interpol[step_index] lowerCAmelCase = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API lowerCAmelCase = 0 lowerCAmelCase = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": lowerCAmelCase = sigma_hat if self.state_in_first_order else sigma_interpol lowerCAmelCase = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": lowerCAmelCase = sigma_hat if self.state_in_first_order else sigma_interpol lowerCAmelCase = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError("""prediction_type not implemented yet: sample""") else: raise ValueError( f"prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`") if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order lowerCAmelCase = (sample - pred_original_sample) / sigma_hat # 3. delta timestep lowerCAmelCase = sigma_interpol - sigma_hat # store for 2nd order step lowerCAmelCase = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order lowerCAmelCase = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep lowerCAmelCase = sigma_next - sigma_hat lowerCAmelCase = self.sample lowerCAmelCase = None lowerCAmelCase = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): """simple docstring""" lowerCAmelCase = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype) if original_samples.device.type == "mps" and torch.is_floating_point(__lowerCAmelCase): # mps does not support float64 lowerCAmelCase = self.timesteps.to(original_samples.device , dtype=torch.floataa) lowerCAmelCase = timesteps.to(original_samples.device , dtype=torch.floataa) else: lowerCAmelCase = self.timesteps.to(original_samples.device) lowerCAmelCase = timesteps.to(original_samples.device) lowerCAmelCase = [self.index_for_timestep(__lowerCAmelCase , __lowerCAmelCase) for t in timesteps] lowerCAmelCase = sigmas[step_indices].flatten() while len(sigma.shape) < len(original_samples.shape): lowerCAmelCase = sigma.unsqueeze(-1) lowerCAmelCase = original_samples + noise * sigma return noisy_samples def __len__( self): """simple docstring""" return self.config.num_train_timesteps
272
'''simple docstring''' import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class a__( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Dict = MvpTokenizer UpperCAmelCase_ : Optional[Any] = MvpTokenizerFast UpperCAmelCase_ : str = True UpperCAmelCase_ : List[Any] = filter_roberta_detectors def a_ ( self): """simple docstring""" super().setUp() lowerCAmelCase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] lowerCAmelCase = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase)))) lowerCAmelCase = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] lowerCAmelCase = {"""unk_token""": """<unk>"""} lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""]) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""]) with open(self.vocab_file , """w""" , encoding="""utf-8""") as fp: fp.write(json.dumps(__lowerCAmelCase) + """\n""") with open(self.merges_file , """w""" , encoding="""utf-8""") as fp: fp.write("""\n""".join(__lowerCAmelCase)) def a_ ( self , **__lowerCAmelCase): """simple docstring""" kwargs.update(self.special_tokens_map) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCAmelCase) def a_ ( self , **__lowerCAmelCase): """simple docstring""" kwargs.update(self.special_tokens_map) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCAmelCase) def a_ ( self , __lowerCAmelCase): """simple docstring""" return "lower newer", "lower newer" @cached_property def a_ ( self): """simple docstring""" return MvpTokenizer.from_pretrained("""RUCAIBox/mvp""") @cached_property def a_ ( self): """simple docstring""" return MvpTokenizerFast.from_pretrained("""RUCAIBox/mvp""") @require_torch def a_ ( self): """simple docstring""" lowerCAmelCase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCAmelCase = [0, 250, 251, 17818, 13, 39186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase = tokenizer(__lowerCAmelCase , max_length=len(__lowerCAmelCase) , padding=__lowerCAmelCase , return_tensors="""pt""") self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase) self.assertEqual((2, 9) , batch.input_ids.shape) self.assertEqual((2, 9) , batch.attention_mask.shape) lowerCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase) # Test that special tokens are reset @require_torch def a_ ( self): """simple docstring""" lowerCAmelCase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase = tokenizer(__lowerCAmelCase , padding=__lowerCAmelCase , return_tensors="""pt""") # check if input_ids are returned and no labels self.assertIn("""input_ids""" , __lowerCAmelCase) self.assertIn("""attention_mask""" , __lowerCAmelCase) self.assertNotIn("""labels""" , __lowerCAmelCase) self.assertNotIn("""decoder_attention_mask""" , __lowerCAmelCase) @require_torch def a_ ( self): """simple docstring""" lowerCAmelCase = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase = tokenizer(text_target=__lowerCAmelCase , max_length=32 , padding="""max_length""" , return_tensors="""pt""") self.assertEqual(32 , targets["""input_ids"""].shape[1]) @require_torch def a_ ( self): """simple docstring""" for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase = tokenizer( ["""I am a small frog""" * 1024, """I am a small frog"""] , padding=__lowerCAmelCase , truncation=__lowerCAmelCase , return_tensors="""pt""") self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase) self.assertEqual(batch.input_ids.shape , (2, 1024)) @require_torch def a_ ( self): """simple docstring""" lowerCAmelCase = ["""A long paragraph for summarization."""] lowerCAmelCase = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase = tokenizer(__lowerCAmelCase , text_target=__lowerCAmelCase , return_tensors="""pt""") lowerCAmelCase = inputs["""input_ids"""] lowerCAmelCase = inputs["""labels"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item()) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item()) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item()) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item()) def a_ ( self): """simple docstring""" pass def a_ ( self): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})"): lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = self.tokenizer_class.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = """A, <mask> AllenNLP sentence.""" lowerCAmelCase = tokenizer_r.encode_plus(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , return_token_type_ids=__lowerCAmelCase) lowerCAmelCase = tokenizer_p.encode_plus(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , return_token_type_ids=__lowerCAmelCase) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""]) , sum(tokens_p["""token_type_ids"""])) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""]) / len(tokens_r["""attention_mask"""]) , sum(tokens_p["""attention_mask"""]) / len(tokens_p["""attention_mask"""]) , ) lowerCAmelCase = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""]) lowerCAmelCase = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""]) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2]) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2]) self.assertSequenceEqual( __lowerCAmelCase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""]) self.assertSequenceEqual( __lowerCAmelCase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""])
272
1
'''simple docstring''' def snake_case__ ( _A: int , _A: list[int] , _A: int ) -> int: '''simple docstring''' def count_of_possible_combinations(_A: int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(_A ) def snake_case__ ( _A: int , _A: list[int] , _A: int ) -> int: '''simple docstring''' def count_of_possible_combinations_with_dp_array( _A: int , _A: list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase = sum( count_of_possible_combinations_with_dp_array(target - item , _A ) for item in array ) lowerCAmelCase = answer return answer lowerCAmelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(_A , _A ) def snake_case__ ( _A: int , _A: list[int] , _A: int ) -> int: '''simple docstring''' lowerCAmelCase = [0] * (target + 1) lowerCAmelCase = 1 for i in range(1 , target + 1 ): for j in range(_A ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __lowercase = 3 __lowercase = 5 __lowercase = [1, 2, 5] print(combination_sum_iv(n, array, target))
272
'''simple docstring''' import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class a__( enum.Enum ): '''simple docstring''' UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : Dict = 1 UpperCAmelCase_ : Any = 2 @add_end_docstrings(lowerCAmelCase__ ) class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : int = ''' In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision and denounces one of the men as a horse thief. Although his father initially slaps him for making such an accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop, begging for his blessing. <eod> </s> <eos> ''' def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" super().__init__(*__lowerCAmelCase , **__lowerCAmelCase) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == """tf""" else MODEL_FOR_CAUSAL_LM_MAPPING) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowerCAmelCase = None if self.model.config.prefix is not None: lowerCAmelCase = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowerCAmelCase = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self._sanitize_parameters(prefix=__lowerCAmelCase , **self._forward_params) lowerCAmelCase = {**self._preprocess_params, **preprocess_params} lowerCAmelCase = {**self._forward_params, **forward_params} def a_ ( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , ): """simple docstring""" lowerCAmelCase = {} if prefix is not None: lowerCAmelCase = prefix if prefix: lowerCAmelCase = self.tokenizer( __lowerCAmelCase , padding=__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , return_tensors=self.framework) lowerCAmelCase = prefix_inputs["""input_ids"""].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f"{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected" """ [None, 'hole']""") lowerCAmelCase = handle_long_generation preprocess_params.update(__lowerCAmelCase) lowerCAmelCase = generate_kwargs lowerCAmelCase = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("""`return_text` is mutually exclusive with `return_full_text`""") if return_tensors is not None: raise ValueError("""`return_full_text` is mutually exclusive with `return_tensors`""") lowerCAmelCase = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("""`return_text` is mutually exclusive with `return_tensors`""") lowerCAmelCase = ReturnType.TENSORS if return_type is not None: lowerCAmelCase = return_type if clean_up_tokenization_spaces is not None: lowerCAmelCase = clean_up_tokenization_spaces if stop_sequence is not None: lowerCAmelCase = self.tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase) if len(__lowerCAmelCase) > 1: warnings.warn( """Stopping on a multiple token sequence is not yet supported on transformers. The first token of""" """ the stop sequence will be used as the stop sequence string in the interim.""") lowerCAmelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def a_ ( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"""add_space_before_punct_symbol""": True}) return super()._parse_and_tokenize(*__lowerCAmelCase , **__lowerCAmelCase) def __call__( self , __lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" return super().__call__(__lowerCAmelCase , **__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase="" , __lowerCAmelCase=None , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.tokenizer( prefix + prompt_text , padding=__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , return_tensors=self.framework) lowerCAmelCase = prompt_text if handle_long_generation == "hole": lowerCAmelCase = inputs["""input_ids"""].shape[-1] if "max_new_tokens" in generate_kwargs: lowerCAmelCase = generate_kwargs["""max_new_tokens"""] else: lowerCAmelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length) - cur_len if new_tokens < 0: raise ValueError("""We cannot infer how many new tokens are expected""") if cur_len + new_tokens > self.tokenizer.model_max_length: lowerCAmelCase = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( """We cannot use `hole` to handle this generation the number of desired tokens exceeds the""" """ models max length""") lowerCAmelCase = inputs["""input_ids"""][:, -keep_length:] if "attention_mask" in inputs: lowerCAmelCase = inputs["""attention_mask"""][:, -keep_length:] return inputs def a_ ( self , __lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = model_inputs["""input_ids"""] lowerCAmelCase = model_inputs.get("""attention_mask""" , __lowerCAmelCase) # Allow empty prompts if input_ids.shape[1] == 0: lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = 1 else: lowerCAmelCase = input_ids.shape[0] lowerCAmelCase = model_inputs.pop("""prompt_text""") # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowerCAmelCase = generate_kwargs.pop("""prefix_length""" , 0) if prefix_length > 0: lowerCAmelCase = """max_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].max_new_tokens is not None ) if not has_max_new_tokens: lowerCAmelCase = generate_kwargs.get("""max_length""") or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowerCAmelCase = """min_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowerCAmelCase = self.model.generate(input_ids=__lowerCAmelCase , attention_mask=__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = generated_sequence.shape[0] if self.framework == "pt": lowerCAmelCase = generated_sequence.reshape(__lowerCAmelCase , out_b // in_b , *generated_sequence.shape[1:]) elif self.framework == "tf": lowerCAmelCase = tf.reshape(__lowerCAmelCase , (in_b, out_b // in_b, *generated_sequence.shape[1:])) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def a_ ( self , __lowerCAmelCase , __lowerCAmelCase=ReturnType.FULL_TEXT , __lowerCAmelCase=True): """simple docstring""" lowerCAmelCase = model_outputs["""generated_sequence"""][0] lowerCAmelCase = model_outputs["""input_ids"""] lowerCAmelCase = model_outputs["""prompt_text"""] lowerCAmelCase = generated_sequence.numpy().tolist() lowerCAmelCase = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowerCAmelCase = {"""generated_token_ids""": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowerCAmelCase = self.tokenizer.decode( __lowerCAmelCase , skip_special_tokens=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowerCAmelCase = 0 else: lowerCAmelCase = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase , )) if return_type == ReturnType.FULL_TEXT: lowerCAmelCase = prompt_text + text[prompt_length:] else: lowerCAmelCase = text[prompt_length:] lowerCAmelCase = {"""generated_text""": all_text} records.append(__lowerCAmelCase) return records
272
1
'''simple docstring''' import cmath import math def snake_case__ ( _A: float , _A: float , _A: float , _A: float ) -> complex: '''simple docstring''' lowerCAmelCase = math.radians(_A ) lowerCAmelCase = math.radians(_A ) # Convert voltage and current to rectangular form lowerCAmelCase = cmath.rect(_A , _A ) lowerCAmelCase = cmath.rect(_A , _A ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
272
'''simple docstring''' def snake_case__ ( _A: str ) -> list[int]: '''simple docstring''' lowerCAmelCase = [0 for i in range(len(_A ) )] # initialize interval's left pointer and right pointer lowerCAmelCase , lowerCAmelCase = 0, 0 for i in range(1 , len(_A ) ): # case when current index is inside the interval if i <= right_pointer: lowerCAmelCase = min(right_pointer - i + 1 , z_result[i - left_pointer] ) lowerCAmelCase = min_edge while go_next(_A , _A , _A ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: lowerCAmelCase , lowerCAmelCase = i, i + z_result[i] - 1 return z_result def snake_case__ ( _A: int , _A: list[int] , _A: str ) -> bool: '''simple docstring''' return i + z_result[i] < len(_A ) and s[z_result[i]] == s[i + z_result[i]] def snake_case__ ( _A: str , _A: str ) -> int: '''simple docstring''' lowerCAmelCase = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string lowerCAmelCase = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(_A ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
272
1
'''simple docstring''' import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __lowercase = logging.get_logger('''transformers.models.speecht5''') __lowercase = { '''speech_encoder_prenet.layer_norm''': '''speecht5.encoder.prenet.feature_projection.layer_norm''', '''speech_encoder_prenet.post_extract_proj''': '''speecht5.encoder.prenet.feature_projection.projection''', '''speech_encoder_prenet.pos_conv.0''': '''speecht5.encoder.prenet.pos_conv_embed.conv''', '''speech_encoder_prenet.mask_emb''': '''speecht5.encoder.prenet.masked_spec_embed''', } __lowercase = { '''text_encoder_prenet.encoder_prenet.0''': '''speecht5.encoder.prenet.embed_tokens''', '''text_encoder_prenet.encoder_prenet.1.alpha''': '''speecht5.encoder.prenet.encode_positions.alpha''', } __lowercase = { '''speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0''': '''speecht5.decoder.prenet.layers.0''', '''speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0''': '''speecht5.decoder.prenet.layers.1''', '''speech_decoder_prenet.decoder_prenet.0.1''': '''speecht5.decoder.prenet.final_layer''', '''speech_decoder_prenet.decoder_prenet.1.alpha''': '''speecht5.decoder.prenet.encode_positions.alpha''', '''speech_decoder_prenet.spkembs_layer.0''': '''speecht5.decoder.prenet.speaker_embeds_layer''', } __lowercase = { '''speech_decoder_postnet.feat_out''': '''speech_decoder_postnet.feat_out''', '''speech_decoder_postnet.prob_out''': '''speech_decoder_postnet.prob_out''', '''speech_decoder_postnet.postnet.postnet.0.0''': '''speech_decoder_postnet.layers.0.conv''', '''speech_decoder_postnet.postnet.postnet.0.1''': '''speech_decoder_postnet.layers.0.batch_norm''', '''speech_decoder_postnet.postnet.postnet.1.0''': '''speech_decoder_postnet.layers.1.conv''', '''speech_decoder_postnet.postnet.postnet.1.1''': '''speech_decoder_postnet.layers.1.batch_norm''', '''speech_decoder_postnet.postnet.postnet.2.0''': '''speech_decoder_postnet.layers.2.conv''', '''speech_decoder_postnet.postnet.postnet.2.1''': '''speech_decoder_postnet.layers.2.batch_norm''', '''speech_decoder_postnet.postnet.postnet.3.0''': '''speech_decoder_postnet.layers.3.conv''', '''speech_decoder_postnet.postnet.postnet.3.1''': '''speech_decoder_postnet.layers.3.batch_norm''', '''speech_decoder_postnet.postnet.postnet.4.0''': '''speech_decoder_postnet.layers.4.conv''', '''speech_decoder_postnet.postnet.postnet.4.1''': '''speech_decoder_postnet.layers.4.batch_norm''', } __lowercase = { '''text_decoder_prenet.embed_tokens''': '''speecht5.decoder.prenet.embed_tokens''', } __lowercase = { '''text_decoder_postnet.output_projection''': '''text_decoder_postnet.lm_head''', } __lowercase = { '''encoder.layers.*.self_attn.k_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj''', '''encoder.layers.*.self_attn.v_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj''', '''encoder.layers.*.self_attn.q_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj''', '''encoder.layers.*.self_attn.out_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj''', '''encoder.layers.*.self_attn_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.layer_norm''', '''encoder.layers.*.fc1''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense''', '''encoder.layers.*.fc2''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense''', '''encoder.layers.*.final_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''speecht5.encoder.wrapped_encoder.layer_norm''', '''encoder.pos_emb.pe_k''': '''speecht5.encoder.wrapped_encoder.embed_positions.pe_k''', } __lowercase = { '''decoder.layers.*.self_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj''', '''decoder.layers.*.self_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj''', '''decoder.layers.*.self_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj''', '''decoder.layers.*.self_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj''', '''decoder.layers.*.self_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm''', '''decoder.layers.*.encoder_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj''', '''decoder.layers.*.encoder_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj''', '''decoder.layers.*.encoder_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj''', '''decoder.layers.*.encoder_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj''', '''decoder.layers.*.encoder_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm''', '''decoder.layers.*.fc1''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense''', '''decoder.layers.*.fc2''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense''', '''decoder.layers.*.final_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm''', } __lowercase = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __lowercase = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __lowercase = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __lowercase = [] __lowercase = [ '''encoder.version''', '''encoder.layers.*.norm_k.weight''', '''encoder.layers.*.norm_k.bias''', '''decoder.version''', '''decoder.layers.*.norm_k.weight''', '''decoder.layers.*.norm_k.bias''', '''decoder.pos_emb.pe_k''', '''speech_encoder_prenet.embed_positions._float_tensor''', '''text_decoder_prenet.embed_positions._float_tensor''', ] __lowercase = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''speech_decoder_prenet.*''', '''speech_decoder_postnet.*''', ] __lowercase = IGNORE_KEYS + [ '''encoder.proj''', '''speech_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] __lowercase = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] def snake_case__ ( _A: List[str] , _A: Union[str, Any] , _A: int , _A: Union[str, Any] , _A: Optional[Any] ) -> Dict: '''simple docstring''' for attribute in key.split(""".""" ): lowerCAmelCase = getattr(_A , _A ) if weight_type is not None: lowerCAmelCase = getattr(_A , _A ).shape else: lowerCAmelCase = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" f" {value.shape} for {full_name}" ) if weight_type == "weight": lowerCAmelCase = value elif weight_type == "weight_g": lowerCAmelCase = value elif weight_type == "weight_v": lowerCAmelCase = value elif weight_type == "bias": lowerCAmelCase = value elif weight_type == "running_mean": lowerCAmelCase = value elif weight_type == "running_var": lowerCAmelCase = value elif weight_type == "num_batches_tracked": lowerCAmelCase = value else: lowerCAmelCase = value logger.info(f"{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}." ) def snake_case__ ( _A: Any , _A: Optional[int] ) -> Optional[Any]: '''simple docstring''' for key in ignore_keys: if key.endswith(""".*""" ): if name.startswith(key[:-1] ): return True elif ".*." in key: lowerCAmelCase , lowerCAmelCase = key.split(""".*.""" ) if prefix in name and suffix in name: return True elif key in name: return True return False def snake_case__ ( _A: Any , _A: Optional[Any] , _A: Tuple ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase = [] if task == "s2t": lowerCAmelCase = hf_model.speechta.encoder.prenet.feature_encoder lowerCAmelCase = MAPPING_S2T lowerCAmelCase = IGNORE_KEYS_S2T elif task == "t2s": lowerCAmelCase = None lowerCAmelCase = MAPPING_T2S lowerCAmelCase = IGNORE_KEYS_T2S elif task == "s2s": lowerCAmelCase = hf_model.speechta.encoder.prenet.feature_encoder lowerCAmelCase = MAPPING_S2S lowerCAmelCase = IGNORE_KEYS_S2S else: raise ValueError(f"Unsupported task: {task}" ) for name, value in fairseq_dict.items(): if should_ignore(_A , _A ): logger.info(f"{name} was ignored" ) continue lowerCAmelCase = False if "conv_layers" in name: load_conv_layer( _A , _A , _A , _A , hf_model.config.feat_extract_norm == """group""" , ) lowerCAmelCase = True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: lowerCAmelCase , lowerCAmelCase = key.split(""".*.""" ) if prefix in name and suffix in name: lowerCAmelCase = suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: lowerCAmelCase = True if "*" in mapped_key: lowerCAmelCase = name.split(_A )[0].split(""".""" )[-2] lowerCAmelCase = mapped_key.replace("""*""" , _A ) if "weight_g" in name: lowerCAmelCase = """weight_g""" elif "weight_v" in name: lowerCAmelCase = """weight_v""" elif "bias" in name: lowerCAmelCase = """bias""" elif "weight" in name: lowerCAmelCase = """weight""" elif "running_mean" in name: lowerCAmelCase = """running_mean""" elif "running_var" in name: lowerCAmelCase = """running_var""" elif "num_batches_tracked" in name: lowerCAmelCase = """num_batches_tracked""" else: lowerCAmelCase = None set_recursively(_A , _A , _A , _A , _A ) continue if not is_used: unused_weights.append(_A ) logger.warning(f"Unused weights: {unused_weights}" ) def snake_case__ ( _A: str , _A: Tuple , _A: Optional[Any] , _A: Tuple , _A: Union[str, Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase = full_name.split("""conv_layers.""" )[-1] lowerCAmelCase = name.split(""".""" ) lowerCAmelCase = int(items[0] ) lowerCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) lowerCAmelCase = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) lowerCAmelCase = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found." ) lowerCAmelCase = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found." ) lowerCAmelCase = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(_A ) @torch.no_grad() def snake_case__ ( _A: Any , _A: Optional[Any] , _A: Tuple , _A: List[Any]=None , _A: int=None , _A: List[str]=None , ) -> Dict: '''simple docstring''' if config_path is not None: lowerCAmelCase = SpeechTaConfig.from_pretrained(_A ) else: lowerCAmelCase = SpeechTaConfig() if task == "s2t": lowerCAmelCase = config.max_text_positions lowerCAmelCase = SpeechTaForSpeechToText(_A ) elif task == "t2s": lowerCAmelCase = 1876 lowerCAmelCase = 600 lowerCAmelCase = config.max_speech_positions lowerCAmelCase = SpeechTaForTextToSpeech(_A ) elif task == "s2s": lowerCAmelCase = 1876 lowerCAmelCase = config.max_speech_positions lowerCAmelCase = SpeechTaForSpeechToSpeech(_A ) else: raise ValueError(f"Unknown task name: {task}" ) if vocab_path: lowerCAmelCase = SpeechTaTokenizer(_A , model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it lowerCAmelCase = AddedToken("""<mask>""" , lstrip=_A , rstrip=_A ) lowerCAmelCase = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) lowerCAmelCase = SpeechTaFeatureExtractor() lowerCAmelCase = SpeechTaProcessor(tokenizer=_A , feature_extractor=_A ) processor.save_pretrained(_A ) lowerCAmelCase = torch.load(_A ) recursively_load_weights(fairseq_checkpoint["""model"""] , _A , _A ) model.save_pretrained(_A ) if repo_id: print("""Pushing to the hub...""" ) processor.push_to_hub(_A ) model.push_to_hub(_A ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument( '''--task''', default='''s2t''', type=str, help='''Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--vocab_path''', default=None, type=str, help='''Path to SentencePiece model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) __lowercase = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
272
'''simple docstring''' from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : str = '''EncodecFeatureExtractor''' UpperCAmelCase_ : Dict = ('''T5Tokenizer''', '''T5TokenizerFast''') def __init__( self , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" super().__init__(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = self.feature_extractor lowerCAmelCase = False def a_ ( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=True): """simple docstring""" return self.tokenizer.get_decoder_prompt_ids(task=__lowerCAmelCase , language=__lowerCAmelCase , no_timestamps=__lowerCAmelCase) def __call__( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" if self._in_target_context_manager: return self.current_processor(*__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = kwargs.pop("""audio""" , __lowerCAmelCase) lowerCAmelCase = kwargs.pop("""sampling_rate""" , __lowerCAmelCase) lowerCAmelCase = kwargs.pop("""text""" , __lowerCAmelCase) if len(__lowerCAmelCase) > 0: lowerCAmelCase = args[0] lowerCAmelCase = args[1:] if audio is None and text is None: raise ValueError("""You need to specify either an `audio` or `text` input to process.""") if text is not None: lowerCAmelCase = self.tokenizer(__lowerCAmelCase , **__lowerCAmelCase) if audio is not None: lowerCAmelCase = self.feature_extractor(__lowerCAmelCase , *__lowerCAmelCase , sampling_rate=__lowerCAmelCase , **__lowerCAmelCase) if audio is None: return inputs elif text is None: return audio_inputs else: lowerCAmelCase = audio_inputs["""input_values"""] if "padding_mask" in audio_inputs: lowerCAmelCase = audio_inputs["""padding_mask"""] return inputs def a_ ( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = kwargs.pop("""audio""" , __lowerCAmelCase) lowerCAmelCase = kwargs.pop("""padding_mask""" , __lowerCAmelCase) if len(__lowerCAmelCase) > 0: lowerCAmelCase = args[0] lowerCAmelCase = args[1:] if audio_values is not None: return self._decode_audio(__lowerCAmelCase , padding_mask=__lowerCAmelCase) else: return self.tokenizer.batch_decode(*__lowerCAmelCase , **__lowerCAmelCase) def a_ ( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" return self.tokenizer.decode(*__lowerCAmelCase , **__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = None): """simple docstring""" lowerCAmelCase = to_numpy(__lowerCAmelCase) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = audio_values.shape if padding_mask is None: return list(__lowerCAmelCase) lowerCAmelCase = to_numpy(__lowerCAmelCase) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) lowerCAmelCase = seq_len - padding_mask.shape[-1] lowerCAmelCase = 1 - self.feature_extractor.padding_value lowerCAmelCase = np.pad(__lowerCAmelCase , ((0, 0), (0, difference)) , """constant""" , constant_values=__lowerCAmelCase) lowerCAmelCase = audio_values.tolist() for i in range(__lowerCAmelCase): lowerCAmelCase = np.asarray(audio_values[i])[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] lowerCAmelCase = sliced_audio.reshape(__lowerCAmelCase , -1) return audio_values
272
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class a__( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Any = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase_ : int = ( { '''feature-extraction''': TFMobileBertModel, '''fill-mask''': TFMobileBertForMaskedLM, '''question-answering''': TFMobileBertForQuestionAnswering, '''text-classification''': TFMobileBertForSequenceClassification, '''token-classification''': TFMobileBertForTokenClassification, '''zero-shot''': TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase_ : int = False UpperCAmelCase_ : Optional[int] = False def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False): """simple docstring""" lowerCAmelCase = super()._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase , return_labels=__lowerCAmelCase) if return_labels: if model_class in get_values(__lowerCAmelCase): lowerCAmelCase = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa) return inputs_dict class a__( lowerCAmelCase__ ): '''simple docstring''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=7 , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=99 , __lowerCAmelCase=32 , __lowerCAmelCase=32 , __lowerCAmelCase=2 , __lowerCAmelCase=4 , __lowerCAmelCase=37 , __lowerCAmelCase="gelu" , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=512 , __lowerCAmelCase=16 , __lowerCAmelCase=2 , __lowerCAmelCase=0.02 , __lowerCAmelCase=3 , __lowerCAmelCase=4 , __lowerCAmelCase=None , ): """simple docstring""" lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_input_mask lowerCAmelCase = use_token_type_ids lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = num_labels lowerCAmelCase = num_choices lowerCAmelCase = scope lowerCAmelCase = embedding_size def a_ ( self): """simple docstring""" lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) lowerCAmelCase = None if self.use_input_mask: lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length]) lowerCAmelCase = None if self.use_token_type_ids: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices) lowerCAmelCase = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = TFMobileBertModel(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) lowerCAmelCase = [input_ids, input_mask] lowerCAmelCase = model(__lowerCAmelCase) lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = TFMobileBertForMaskedLM(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = TFMobileBertForNextSentencePrediction(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = TFMobileBertForPreTraining(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.num_labels lowerCAmelCase = TFMobileBertForSequenceClassification(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.num_choices lowerCAmelCase = TFMobileBertForMultipleChoice(config=__lowerCAmelCase) lowerCAmelCase = tf.tile(tf.expand_dims(__lowerCAmelCase , 1) , (1, self.num_choices, 1)) lowerCAmelCase = tf.tile(tf.expand_dims(__lowerCAmelCase , 1) , (1, self.num_choices, 1)) lowerCAmelCase = tf.tile(tf.expand_dims(__lowerCAmelCase , 1) , (1, self.num_choices, 1)) lowerCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.num_labels lowerCAmelCase = TFMobileBertForTokenClassification(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = TFMobileBertForQuestionAnswering(config=__lowerCAmelCase) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = model(__lowerCAmelCase) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def a_ ( self): """simple docstring""" lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict def a_ ( self): """simple docstring""" lowerCAmelCase = TFMobileBertModelTest.TFMobileBertModelTester(self) lowerCAmelCase = ConfigTester(self , config_class=__lowerCAmelCase , hidden_size=37) def a_ ( self): """simple docstring""" self.config_tester.run_common_tests() def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*__lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*__lowerCAmelCase) @slow def a_ ( self): """simple docstring""" for model_name in ["google/mobilebert-uncased"]: lowerCAmelCase = TFMobileBertModel.from_pretrained(__lowerCAmelCase) self.assertIsNotNone(__lowerCAmelCase) @require_tf class a__( unittest.TestCase ): '''simple docstring''' @slow def a_ ( self): """simple docstring""" lowerCAmelCase = TFMobileBertForPreTraining.from_pretrained("""google/mobilebert-uncased""") lowerCAmelCase = tf.constant([[0, 1, 2, 3, 4, 5]]) lowerCAmelCase = model(__lowerCAmelCase)[0] lowerCAmelCase = [1, 6, 30522] self.assertEqual(output.shape , __lowerCAmelCase) lowerCAmelCase = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ]) tf.debugging.assert_near(output[:, :3, :3] , __lowerCAmelCase , atol=1E-4)
272
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class a__( unittest.TestCase ): '''simple docstring''' @property def a_ ( self): """simple docstring""" torch.manual_seed(0) lowerCAmelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def a_ ( self): """simple docstring""" lowerCAmelCase = self.dummy_uncond_unet lowerCAmelCase = PNDMScheduler() lowerCAmelCase = PNDMPipeline(unet=__lowerCAmelCase , scheduler=__lowerCAmelCase) pndm.to(__lowerCAmelCase) pndm.set_progress_bar_config(disable=__lowerCAmelCase) lowerCAmelCase = torch.manual_seed(0) lowerCAmelCase = pndm(generator=__lowerCAmelCase , num_inference_steps=20 , output_type="""numpy""").images lowerCAmelCase = torch.manual_seed(0) lowerCAmelCase = pndm(generator=__lowerCAmelCase , num_inference_steps=20 , output_type="""numpy""" , return_dict=__lowerCAmelCase)[0] lowerCAmelCase = image[0, -3:, -3:, -1] lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 @slow @require_torch class a__( unittest.TestCase ): '''simple docstring''' def a_ ( self): """simple docstring""" lowerCAmelCase = """google/ddpm-cifar10-32""" lowerCAmelCase = UNetaDModel.from_pretrained(__lowerCAmelCase) lowerCAmelCase = PNDMScheduler() lowerCAmelCase = PNDMPipeline(unet=__lowerCAmelCase , scheduler=__lowerCAmelCase) pndm.to(__lowerCAmelCase) pndm.set_progress_bar_config(disable=__lowerCAmelCase) lowerCAmelCase = torch.manual_seed(0) lowerCAmelCase = pndm(generator=__lowerCAmelCase , output_type="""numpy""").images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase = np.array([0.1564, 0.14645, 0.1406, 0.14715, 0.12425, 0.14045, 0.13115, 0.12175, 0.125]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2
272
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''transfo-xl-wt103''': '''https://huggingface.co/transfo-xl-wt103/resolve/main/config.json''', } class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : List[str] = '''transfo-xl''' UpperCAmelCase_ : Any = ['''mems'''] UpperCAmelCase_ : List[Any] = { '''n_token''': '''vocab_size''', '''hidden_size''': '''d_model''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self , __lowerCAmelCase=267735 , __lowerCAmelCase=[20000, 40000, 200000] , __lowerCAmelCase=1024 , __lowerCAmelCase=1024 , __lowerCAmelCase=16 , __lowerCAmelCase=64 , __lowerCAmelCase=4096 , __lowerCAmelCase=4 , __lowerCAmelCase=False , __lowerCAmelCase=18 , __lowerCAmelCase=1600 , __lowerCAmelCase=1000 , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=0 , __lowerCAmelCase=-1 , __lowerCAmelCase=True , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.0 , __lowerCAmelCase=True , __lowerCAmelCase="normal" , __lowerCAmelCase=0.01 , __lowerCAmelCase=0.01 , __lowerCAmelCase=0.02 , __lowerCAmelCase=1E-5 , __lowerCAmelCase=0 , **__lowerCAmelCase , ): """simple docstring""" lowerCAmelCase = vocab_size lowerCAmelCase = [] self.cutoffs.extend(__lowerCAmelCase) if proj_share_all_but_first: lowerCAmelCase = [False] + [True] * len(self.cutoffs) else: lowerCAmelCase = [False] + [False] * len(self.cutoffs) lowerCAmelCase = d_model lowerCAmelCase = d_embed lowerCAmelCase = d_head lowerCAmelCase = d_inner lowerCAmelCase = div_val lowerCAmelCase = pre_lnorm lowerCAmelCase = n_layer lowerCAmelCase = n_head lowerCAmelCase = mem_len lowerCAmelCase = same_length lowerCAmelCase = attn_type lowerCAmelCase = clamp_len lowerCAmelCase = sample_softmax lowerCAmelCase = adaptive lowerCAmelCase = dropout lowerCAmelCase = dropatt lowerCAmelCase = untie_r lowerCAmelCase = init lowerCAmelCase = init_range lowerCAmelCase = proj_init_std lowerCAmelCase = init_std lowerCAmelCase = layer_norm_epsilon super().__init__(eos_token_id=__lowerCAmelCase , **__lowerCAmelCase) @property def a_ ( self): """simple docstring""" logger.info(f"The model {self.model_type} is one of the few models that has no sequence length limit.") return -1 @max_position_embeddings.setter def a_ ( self , __lowerCAmelCase): """simple docstring""" raise NotImplementedError( f"The model {self.model_type} is one of the few models that has no sequence length limit.")
272
'''simple docstring''' from string import ascii_lowercase, ascii_uppercase def snake_case__ ( _A: str ) -> str: '''simple docstring''' if not sentence: return "" lowerCAmelCase = dict(zip(_A , _A ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
272
1