code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from typing import Any import numpy as np def __lowerCamelCase ( __snake_case : np.ndarray ) -> List[Any]: """simple docstring""" return np.array_equal(snake_case__, matrix.conjugate().T ) def __lowerCamelCase ( __snake_case : np.ndarray, __snake_case : np.ndarray ) -> Optional[int]: """simple docstring""" A__ : Optional[int] =v.conjugate().T A__ : List[Any] =v_star.dot(snake_case__ ) assert isinstance(snake_case__, np.ndarray ) return (v_star_dot.dot(snake_case__ )) / (v_star.dot(snake_case__ )) def __lowerCamelCase ( ) -> Optional[Any]: """simple docstring""" A__ : int =np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]] ) A__ : int =np.array([[1], [2], [3]] ) assert is_hermitian(snake_case__ ), f"{a} is not hermitian." print(rayleigh_quotient(snake_case__, snake_case__ ) ) A__ : int =np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]] ) assert is_hermitian(snake_case__ ), f"{a} is not hermitian." assert rayleigh_quotient(snake_case__, snake_case__ ) == float(3 ) if __name__ == "__main__": import doctest doctest.testmod() tests()
134
"""simple docstring""" import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowercase( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Any = ort.SessionOptions() _snake_case : Union[str, Any] = False return options def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) _snake_case : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) _snake_case : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy""" ) # using the PNDM scheduler by default _snake_case : Optional[Any] = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( """CompVis/stable-diffusion-v1-4""", revision="""onnx""", safety_checker=a_, feature_extractor=a_, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=a_ ) _snake_case : Optional[Any] = """A red cat sitting on a park bench""" _snake_case : Optional[int] = np.random.RandomState(0 ) _snake_case : Any = pipe( prompt=a_, image=a_, mask_image=a_, strength=0.75, guidance_scale=7.5, num_inference_steps=15, generator=a_, output_type="""np""", ) _snake_case : Dict = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
64
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[Any] = { '''facebook/deit-base-distilled-patch16-224''': ( '''https://huggingface.co/facebook/deit-base-patch16-224/resolve/main/config.json''' ), # See all DeiT models at https://huggingface.co/models?filter=deit } class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = "deit" def __init__( self : Tuple ,lowercase_ : Dict=7_6_8 ,lowercase_ : int=1_2 ,lowercase_ : Tuple=1_2 ,lowercase_ : Tuple=3_0_7_2 ,lowercase_ : List[Any]="gelu" ,lowercase_ : Optional[int]=0.0 ,lowercase_ : Dict=0.0 ,lowercase_ : Optional[int]=0.02 ,lowercase_ : Tuple=1E-12 ,lowercase_ : Union[str, Any]=2_2_4 ,lowercase_ : Dict=1_6 ,lowercase_ : int=3 ,lowercase_ : Union[str, Any]=True ,lowercase_ : Optional[Any]=1_6 ,**lowercase_ : int ,): super().__init__(**lowercase_ ) lowerCAmelCase__ : Any = hidden_size lowerCAmelCase__ : int = num_hidden_layers lowerCAmelCase__ : List[Any] = num_attention_heads lowerCAmelCase__ : Dict = intermediate_size lowerCAmelCase__ : Tuple = hidden_act lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : int = attention_probs_dropout_prob lowerCAmelCase__ : Any = initializer_range lowerCAmelCase__ : str = layer_norm_eps lowerCAmelCase__ : Tuple = image_size lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = num_channels lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : int = encoder_stride class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = version.parse("1.11" ) @property def __lowerCAmelCase ( self : str ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def __lowerCAmelCase ( self : Any ): return 1E-4
74
"""simple docstring""" __UpperCamelCase : Optional[Any] = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __SCREAMING_SNAKE_CASE ( A_ ): # Make sure the supplied data is a bytes-like object if not isinstance(A_ , A_ ): lowerCAmelCase__ : Dict = f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(A_ ) lowerCAmelCase__ : Any = ''''''.join(bin(A_ )[2:].zfill(8 ) for byte in data ) lowerCAmelCase__ : List[str] = len(A_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowerCAmelCase__ : List[str] = b'''=''' * ((6 - len(A_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(A_ ) % 6) else: lowerCAmelCase__ : Tuple = b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(A_ ) , 6 ) ).encode() + padding ) def __SCREAMING_SNAKE_CASE ( A_ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(A_ , A_ ) and not isinstance(A_ , A_ ): lowerCAmelCase__ : str = ( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(A_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(A_ , A_ ): try: lowerCAmelCase__ : Union[str, Any] = encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowerCAmelCase__ : Union[str, Any] = encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(A_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowerCAmelCase__ : List[str] = encoded_data[:-padding] lowerCAmelCase__ : Tuple = ''''''.join( bin(B64_CHARSET.index(A_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowerCAmelCase__ : Any = ''''''.join( bin(B64_CHARSET.index(A_ ) )[2:].zfill(6 ) for char in encoded_data ) lowerCAmelCase__ : List[Any] = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(A_ ) , 8 ) ] return bytes(A_ ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
'''simple docstring''' import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class __UpperCamelCase ( lowerCAmelCase_ ): def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=False , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=64 , __a="gelu" , __a=0.1 , __a=0.1 , __a=512 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , __a=2 , __a=2 , __a=2 , __a=2 , __a=4 , __a=1 , ): '''simple docstring''' __a : Union[str, Any] = parent __a : Union[str, Any] = batch_size __a : str = seq_length __a : str = is_training __a : str = use_input_mask __a : Dict = use_token_type_ids __a : List[str] = use_labels __a : Union[str, Any] = vocab_size __a : int = hidden_size __a : Any = num_hidden_layers __a : Dict = num_attention_heads __a : str = intermediate_size __a : Optional[int] = hidden_act __a : int = hidden_dropout_prob __a : str = attention_probs_dropout_prob __a : List[str] = max_position_embeddings __a : List[str] = type_vocab_size __a : Tuple = type_sequence_label_size __a : Optional[Any] = initializer_range __a : Tuple = num_labels __a : Tuple = num_choices __a : Dict = scope __a : int = q_groups __a : Optional[Any] = k_groups __a : List[str] = v_groups __a : Union[str, Any] = post_attention_groups __a : List[Any] = intermediate_groups __a : Dict = output_groups def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a : str = None if self.use_input_mask: __a : int = random_attention_mask([self.batch_size, self.seq_length] ) __a : Tuple = None __a : List[str] = None __a : Tuple = None if self.use_labels: __a : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __a : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) __a : Optional[int] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : List[str] = SqueezeBertModel(config=__a ) model.to(__a ) model.eval() __a : List[Any] = model(__a , __a ) __a : List[str] = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Tuple = SqueezeBertForMaskedLM(config=__a ) model.to(__a ) model.eval() __a : Dict = model(__a , attention_mask=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Optional[int] = SqueezeBertForQuestionAnswering(config=__a ) model.to(__a ) model.eval() __a : int = model( __a , attention_mask=__a , start_positions=__a , end_positions=__a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : List[Any] = self.num_labels __a : int = SqueezeBertForSequenceClassification(__a ) model.to(__a ) model.eval() __a : List[Any] = model(__a , attention_mask=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Optional[Any] = self.num_labels __a : Dict = SqueezeBertForTokenClassification(config=__a ) model.to(__a ) model.eval() __a : List[Any] = model(__a , attention_mask=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : str = self.num_choices __a : int = SqueezeBertForMultipleChoice(config=__a ) model.to(__a ) model.eval() __a : Union[str, Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : Tuple = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : Dict = model( __a , attention_mask=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = self.prepare_config_and_inputs() ((__a) , (__a) , (__a) , (__a) , (__a) , (__a)) : Any = config_and_inputs __a : Optional[int] = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): A_ = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) A_ = ( { "feature-extraction": SqueezeBertModel, "fill-mask": SqueezeBertForMaskedLM, "question-answering": SqueezeBertForQuestionAnswering, "text-classification": SqueezeBertForSequenceClassification, "token-classification": SqueezeBertForTokenClassification, "zero-shot": SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) A_ = False A_ = True A_ = False def __UpperCAmelCase ( self ): '''simple docstring''' __a : Any = SqueezeBertModelTester(self ) __a : Optional[int] = ConfigTester(self , config_class=__a , dim=37 ) def __UpperCAmelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*__a ) @slow def __UpperCAmelCase ( self ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : int = SqueezeBertModel.from_pretrained(__a ) self.assertIsNotNone(__a ) @require_sentencepiece @require_tokenizers @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) __a : Optional[int] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]] ) __a : List[Any] = model(__a )[0] __a : int = torch.Size((1, 3) ) self.assertEqual(output.shape , __a ) __a : str = torch.tensor([[0.6401, -0.0349, -0.6041]] ) self.assertTrue(torch.allclose(__a , __a , atol=1E-4 ) )
27
from __future__ import annotations import os from collections.abc import Mapping a_ = tuple[int, int] class lowercase__ : def __init__( self , __UpperCAmelCase , __UpperCAmelCase )-> None: '''simple docstring''' lowerCAmelCase__ = vertices lowerCAmelCase__ = { (min(__UpperCAmelCase ), max(__UpperCAmelCase )): weight for edge, weight in edges.items() } def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase )-> None: '''simple docstring''' self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) lowerCAmelCase__ = weight def UpperCAmelCase ( self )-> Graph: '''simple docstring''' lowerCAmelCase__ = Graph({min(self.vertices )} , {} ) lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 while len(subgraph.vertices ) < len(self.vertices ): lowerCAmelCase__ = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: lowerCAmelCase__ = edge lowerCAmelCase__ = weight subgraph.add_edge(__UpperCAmelCase , __UpperCAmelCase ) return subgraph def _a ( UpperCamelCase_ : str = "p107_network.txt" ) -> int: """simple docstring""" lowerCAmelCase__ = os.path.abspath(os.path.dirname(UpperCamelCase_ ) ) lowerCAmelCase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) lowerCAmelCase__ = {} lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 with open(UpperCamelCase_ ) as f: lowerCAmelCase__ = f.read().strip().split("\n" ) lowerCAmelCase__ = [line.split("," ) for line in data] for edgea in range(1 , len(UpperCamelCase_ ) ): for edgea in range(UpperCamelCase_ ): if adjaceny_matrix[edgea][edgea] != "-": lowerCAmelCase__ = int(adjaceny_matrix[edgea][edgea] ) lowerCAmelCase__ = Graph(set(range(len(UpperCamelCase_ ) ) ) , UpperCamelCase_ ) lowerCAmelCase__ = graph.prims_algorithm() lowerCAmelCase__ = sum(graph.edges.values() ) lowerCAmelCase__ = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"{solution() = }")
340
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def A_ ( A__ , A__=False ) -> str: a__ : Any = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'module.blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'module.blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append( (F'module.blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'module.blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'module.blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'module.blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'module.blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'module.blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'module.blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'module.blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ('module.cls_token', 'vit.embeddings.cls_token'), ('module.patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('module.patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('module.pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('module.norm.weight', 'layernorm.weight'), ('module.norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" a__ : Tuple = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def A_ ( A__ , A__ , A__=False ) -> Tuple: for i in range(config.num_hidden_layers ): if base_model: a__ : List[str] = '' else: a__ : Tuple = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) a__ : List[str] = state_dict.pop(F'module.blocks.{i}.attn.qkv.weight' ) a__ : Optional[Any] = state_dict.pop(F'module.blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict a__ : List[str] = in_proj_weight[ : config.hidden_size, : ] a__ : List[Any] = in_proj_bias[: config.hidden_size] a__ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] a__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] a__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] a__ : str = in_proj_bias[-config.hidden_size :] def A_ ( A__ ) -> Optional[int]: a__ : List[str] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(A__ , A__ ) def A_ ( A__ ) -> Optional[int]: # projection head is used in the self-supervised pre-training in MSN, # for downstream task it's not needed. a__ : Dict = [ 'module.fc.fc1.weight', 'module.fc.fc1.bias', 'module.fc.bn1.weight', 'module.fc.bn1.bias', 'module.fc.bn1.running_mean', 'module.fc.bn1.running_var', 'module.fc.bn1.num_batches_tracked', 'module.fc.fc2.weight', 'module.fc.fc2.bias', 'module.fc.bn2.weight', 'module.fc.bn2.bias', 'module.fc.bn2.running_mean', 'module.fc.bn2.running_var', 'module.fc.bn2.num_batches_tracked', 'module.fc.fc3.weight', 'module.fc.fc3.bias', ] for k in ignore_keys: state_dict.pop(A__ , A__ ) def A_ ( A__ , A__ , A__ ) -> int: a__ : List[Any] = dct.pop(A__ ) a__ : Union[str, Any] = val def A_ ( A__ , A__ ) -> Dict: a__ : Tuple = ViTMSNConfig() a__ : Dict = 1000 a__ : Optional[Any] = 'datasets/huggingface/label-files' a__ : str = 'imagenet-1k-id2label.json' a__ : Optional[Any] = json.load(open(hf_hub_download(A__ , A__ ) , 'r' ) ) a__ : Tuple = {int(A__ ): v for k, v in idalabel.items()} a__ : str = idalabel a__ : List[str] = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: a__ : int = 384 a__ : Optional[int] = 1536 a__ : int = 6 elif "l16" in checkpoint_url: a__ : str = 1024 a__ : Any = 4096 a__ : str = 24 a__ : Dict = 16 a__ : Optional[int] = 0.1 elif "b4" in checkpoint_url: a__ : Tuple = 4 elif "l7" in checkpoint_url: a__ : Union[str, Any] = 7 a__ : Union[str, Any] = 1024 a__ : Dict = 4096 a__ : Dict = 24 a__ : str = 16 a__ : Tuple = 0.1 a__ : Dict = ViTMSNModel(A__ ) a__ : Optional[Any] = torch.hub.load_state_dict_from_url(A__ , map_location='cpu' )['target_encoder'] a__ : str = ViTImageProcessor(size=config.image_size ) remove_projection_head(A__ ) a__ : str = create_rename_keys(A__ , base_model=A__ ) for src, dest in rename_keys: rename_key(A__ , A__ , A__ ) read_in_q_k_v(A__ , A__ , base_model=A__ ) model.load_state_dict(A__ ) model.eval() a__ : int = 'http://images.cocodataset.org/val2017/000000039769.jpg' a__ : Tuple = Image.open(requests.get(A__ , stream=A__ ).raw ) a__ : Optional[int] = ViTImageProcessor( size=config.image_size , image_mean=A__ , image_std=A__ ) a__ : Optional[Any] = image_processor(images=A__ , return_tensors='pt' ) # forward pass torch.manual_seed(2 ) a__ : str = model(**A__ ) a__ : int = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: a__ : List[Any] = torch.tensor([[-1.09_15, -1.48_76, -1.18_09]] ) elif "b16" in checkpoint_url: a__ : List[Any] = torch.tensor([[14.28_89, -18.90_45, 11.72_81]] ) elif "l16" in checkpoint_url: a__ : Tuple = torch.tensor([[41.50_28, -22.86_81, 45.64_75]] ) elif "b4" in checkpoint_url: a__ : Dict = torch.tensor([[-4.38_68, 5.29_32, -0.41_37]] ) else: a__ : Optional[Any] = torch.tensor([[-0.17_92, -0.64_65, 2.42_63]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , A__ , atol=1E-4 ) print(F'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(A__ ) if __name__ == "__main__": lowercase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar""", type=str, help="""URL of the checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowercase : str = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
225
from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def A_ ( A__ ) -> Dict: if isinstance(A__ , collections.abc.Iterable ): return x return (x, x) @require_tf class A__ : """simple docstring""" def __lowercase ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' pass def __lowercase ( self) -> Dict: '''simple docstring''' pass def __lowercase ( self) -> Dict: '''simple docstring''' pass def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase) -> List[Any]: '''simple docstring''' a__ : Any = VisionTextDualEncoderConfig.from_vision_text_configs(lowercase , lowercase) a__ : Any = TFVisionTextDualEncoderModel(lowercase) a__ : Dict = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], config.projection_dim)) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], config.projection_dim)) def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase) -> List[Any]: '''simple docstring''' a__ , a__ : List[Any] = self.get_vision_text_model(lowercase , lowercase) a__ : List[str] = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase) a__ : Dict = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], model.config.projection_dim)) def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' a__ , a__ : Any = self.get_vision_text_model(lowercase , lowercase) a__ : Tuple = {'vision_model': vision_model, 'text_model': text_model} a__ : Union[str, Any] = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**lowercase) a__ : Any = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], model.config.projection_dim)) def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase) -> Optional[Any]: '''simple docstring''' a__ , a__ : int = self.get_vision_text_model(lowercase , lowercase) a__ : List[Any] = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase) a__ : Optional[Any] = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase) a__ : int = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase) a__ : str = TFVisionTextDualEncoderModel.from_pretrained(lowercase) a__ : List[str] = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase) a__ : str = after_output[0].numpy() a__ : str = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(lowercase , 1e-5) def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase) -> Optional[int]: '''simple docstring''' a__ , a__ : Optional[Any] = self.get_vision_text_model(lowercase , lowercase) a__ : Dict = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase) a__ : Optional[int] = model( input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase , output_attentions=lowercase) a__ : List[str] = output.vision_model_output.attentions self.assertEqual(len(lowercase) , vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) a__ : Optional[Any] = to_atuple(vision_model.config.image_size) a__ : Dict = to_atuple(vision_model.config.patch_size) a__ : int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) a__ : List[str] = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) a__ : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(lowercase) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def __lowercase ( self , lowercase , lowercase , lowercase) -> Any: '''simple docstring''' a__ : str = np.abs((a - b)).max() self.assertLessEqual(lowercase , lowercase , F'Difference between torch and flax is {diff} (>= {tol}).') def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**lowercase) def __lowercase ( self) -> Dict: '''simple docstring''' a__ : List[Any] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowercase) def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : Any = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowercase) def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : int = self.prepare_config_and_inputs() self.check_save_load(**lowercase) def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowercase) @slow def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ , a__ : Union[str, Any] = self.get_pretrained_model_and_inputs() a__ : Optional[int] = model_a(**lowercase) a__ : int = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowercase) a__ : Union[str, Any] = TFVisionTextDualEncoderModel.from_pretrained(lowercase) a__ : int = model_a(**lowercase) a__ : str = after_outputs[0].numpy() a__ : List[Any] = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(lowercase , 1e-5) @require_tf class A__ ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : Any = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-vit' , 'hf-internal-testing/tiny-random-bert') a__ : str = 13 a__ : Optional[int] = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ]) a__ : Tuple = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size) a__ : Optional[int] = random_attention_mask([batch_size, 4]) a__ : str = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def __lowercase ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__ : Optional[Any] = TFViTModel(lowercase , name='vision_model') a__ : Tuple = TFBertModel(lowercase , name='text_model') return vision_model, text_model def __lowercase ( self) -> Any: '''simple docstring''' a__ : Tuple = TFViTModelTester(self) a__ : int = TFBertModelTester(self) a__ : Optional[Any] = vit_model_tester.prepare_config_and_inputs() a__ : Any = bert_model_tester.prepare_config_and_inputs() a__ , a__ , a__ : Optional[int] = vision_config_and_inputs ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ) : Any = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class A__ ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" def __lowercase ( self) -> List[str]: '''simple docstring''' a__ : Optional[Any] = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'Rocketknight1/tiny-random-deit-tf' , 'hf-internal-testing/tiny-random-roberta') a__ : Union[str, Any] = 13 a__ : Dict = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ]) a__ : Optional[Any] = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size) a__ : Any = random_attention_mask([batch_size, 4]) a__ : int = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase) -> Optional[Any]: '''simple docstring''' a__ , a__ : List[Any] = self.get_vision_text_model(lowercase , lowercase) a__ : Any = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase) a__ : int = model( input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase , output_attentions=lowercase) a__ : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(lowercase) , vision_config.num_hidden_layers) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) a__ : Optional[int] = to_atuple(vision_model.config.image_size) a__ : str = to_atuple(vision_model.config.patch_size) a__ : Tuple = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) a__ : List[str] = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) a__ : List[Any] = output.text_model_output.attentions self.assertEqual(len(lowercase) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def __lowercase ( self , lowercase , lowercase) -> Optional[Any]: '''simple docstring''' a__ : List[str] = TFDeiTModel(lowercase , name='vision_model') a__ : Optional[int] = TFRobertaModel(lowercase , name='text_model') return vision_model, text_model def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : Optional[int] = TFDeiTModelTester(self) a__ : str = TFRobertaModelTester(self) a__ : str = vit_model_tester.prepare_config_and_inputs() a__ : Dict = bert_model_tester.prepare_config_and_inputs() a__ , a__ , a__ : Any = vision_config_and_inputs ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ) : Tuple = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class A__ ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" def __lowercase ( self) -> Any: '''simple docstring''' a__ : Dict = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'Rocketknight1/tiny-random-clip-tf' , 'hf-internal-testing/tiny-random-bert') a__ : Optional[int] = 13 a__ : Optional[Any] = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ]) a__ : int = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size) a__ : str = random_attention_mask([batch_size, 4]) a__ : int = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def __lowercase ( self , lowercase , lowercase) -> Optional[Any]: '''simple docstring''' a__ : str = TFCLIPVisionModel(lowercase , name='vision_model') a__ : str = TFBertModel(lowercase , name='text_model') return vision_model, text_model def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : List[str] = TFCLIPVisionModelTester(self) a__ : Dict = TFBertModelTester(self) a__ : Optional[Any] = clip_model_tester.prepare_config_and_inputs() a__ : List[Any] = bert_model_tester.prepare_config_and_inputs() a__ , a__ : Union[str, Any] = vision_config_and_inputs ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ) : Optional[int] = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class A__ ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self) -> Any: '''simple docstring''' a__ : Tuple = TFVisionTextDualEncoderModel.from_pretrained( 'clip-italian/clip-italian' , logit_scale_init_value=1.0 , from_pt=lowercase) a__ : str = VisionTextDualEncoderProcessor.from_pretrained('clip-italian/clip-italian') a__ : Tuple = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png') a__ : Optional[Any] = processor( text=['una foto di un gatto', 'una foto di un cane'] , images=lowercase , padding=lowercase , return_tensors='np') a__ : int = model(**lowercase) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) a__ : List[str] = np.array([[1.2_28_47_27, 0.3_10_41_22]]) self.assertTrue(np.allclose(outputs.logits_per_image.numpy() , lowercase , atol=1e-3))
225
1
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html UpperCamelCase = '''platform''' import jax import jax.numpy as jnp from transformers.models.blenderbot_small.modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, shift_tokens_right, ) def lowercase_ ( _lowerCamelCase : int , _lowerCamelCase : str , _lowerCamelCase : str=None , _lowerCamelCase : int=None , _lowerCamelCase : str=None , _lowerCamelCase : Optional[Any]=None , _lowerCamelCase : str=None , _lowerCamelCase : str=None , ): if attention_mask is None: lowercase__ : Dict = np.where(input_ids != config.pad_token_id , 1 , 0) if decoder_attention_mask is None: lowercase__ : Optional[Any] = np.where(decoder_input_ids != config.pad_token_id , 1 , 0) if head_mask is None: lowercase__ : str = np.ones((config.encoder_layers, config.encoder_attention_heads)) if decoder_head_mask is None: lowercase__ : Dict = np.ones((config.decoder_layers, config.decoder_attention_heads)) if cross_attn_head_mask is None: lowercase__ : int = np.ones((config.decoder_layers, config.decoder_attention_heads)) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class snake_case_ : def __init__( self : Dict , lowercase_ : Any , lowercase_ : List[str]=13 , lowercase_ : List[Any]=7 , lowercase_ : int=True , lowercase_ : Any=False , lowercase_ : int=99 , lowercase_ : Tuple=16 , lowercase_ : Optional[Any]=2 , lowercase_ : Union[str, Any]=4 , lowercase_ : Dict=4 , lowercase_ : Optional[Any]="gelu" , lowercase_ : Optional[int]=0.1 , lowercase_ : Tuple=0.1 , lowercase_ : Optional[Any]=32 , lowercase_ : Tuple=2 , lowercase_ : Any=1 , lowercase_ : List[str]=0 , lowercase_ : Union[str, Any]=0.02 , ) -> Dict: lowercase__ : Optional[int] = parent lowercase__ : str = batch_size lowercase__ : str = seq_length lowercase__ : Tuple = is_training lowercase__ : Optional[Any] = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : Dict = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Dict = intermediate_size lowercase__ : Tuple = hidden_act lowercase__ : List[str] = hidden_dropout_prob lowercase__ : Union[str, Any] = attention_probs_dropout_prob lowercase__ : List[Any] = max_position_embeddings lowercase__ : Optional[int] = eos_token_id lowercase__ : List[Any] = pad_token_id lowercase__ : Union[str, Any] = bos_token_id lowercase__ : Optional[int] = initializer_range def __UpperCamelCase ( self : Optional[Any] ) -> Tuple: lowercase__ : Any = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) lowercase__ : Tuple = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) lowercase__ : Tuple = shift_tokens_right(lowercase_ , 1 , 2 ) lowercase__ : Union[str, Any] = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowercase_ , ) lowercase__ : List[Any] = prepare_blenderbot_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) return config, inputs_dict def __UpperCamelCase ( self : Tuple ) -> int: lowercase__ , lowercase__ : List[Any] = self.prepare_config_and_inputs() return config, inputs_dict def __UpperCamelCase ( self : Any , lowercase_ : List[Any] , lowercase_ : Any , lowercase_ : List[Any] ) -> List[Any]: lowercase__ : Union[str, Any] = 20 lowercase__ : List[Any] = model_class_name(lowercase_ ) lowercase__ : Optional[int] = model.encode(inputs_dict["input_ids"] ) lowercase__ , lowercase__ : Optional[int] = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) lowercase__ : Optional[int] = model.init_cache(decoder_input_ids.shape[0] , lowercase_ , lowercase_ ) lowercase__ : Tuple = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) lowercase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase__ : int = model.decode( decoder_input_ids[:, :-1] , lowercase_ , decoder_attention_mask=lowercase_ , past_key_values=lowercase_ , decoder_position_ids=lowercase_ , ) lowercase__ : str = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) lowercase__ : Dict = model.decode( decoder_input_ids[:, -1:] , lowercase_ , decoder_attention_mask=lowercase_ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowercase_ , ) lowercase__ : Union[str, Any] = model.decode(lowercase_ , lowercase_ ) lowercase__ : Tuple = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) def __UpperCamelCase ( self : Any , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Any ) -> Union[str, Any]: lowercase__ : Optional[Any] = 20 lowercase__ : List[str] = model_class_name(lowercase_ ) lowercase__ : Any = model.encode(inputs_dict["input_ids"] ) lowercase__ , lowercase__ : List[str] = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) lowercase__ : Tuple = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) lowercase__ : List[str] = model.init_cache(decoder_input_ids.shape[0] , lowercase_ , lowercase_ ) lowercase__ : Tuple = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase__ : str = model.decode( decoder_input_ids[:, :-1] , lowercase_ , decoder_attention_mask=lowercase_ , past_key_values=lowercase_ , decoder_position_ids=lowercase_ , ) lowercase__ : Union[str, Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) lowercase__ : int = model.decode( decoder_input_ids[:, -1:] , lowercase_ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowercase_ , decoder_position_ids=lowercase_ , ) lowercase__ : int = model.decode(lowercase_ , lowercase_ , decoder_attention_mask=lowercase_ ) lowercase__ : Dict = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) @require_flax class snake_case_ ( unittest.TestCase ): __A : Dict = 99 def __UpperCamelCase ( self : Any ) -> Tuple: lowercase__ : Tuple = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) lowercase__ : Tuple = input_ids.shape[0] lowercase__ : Union[str, Any] = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def __UpperCamelCase ( self : str ) -> Tuple: lowercase__ , lowercase__ , lowercase__ : Optional[int] = self._get_config_and_data() lowercase__ : Tuple = FlaxBlenderbotSmallForConditionalGeneration(lowercase_ ) lowercase__ : List[str] = lm_model(input_ids=lowercase_ ) lowercase__ : int = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["logits"].shape , lowercase_ ) def __UpperCamelCase ( self : Optional[Any] ) -> str: lowercase__ : List[Any] = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) lowercase__ : Dict = FlaxBlenderbotSmallForConditionalGeneration(lowercase_ ) lowercase__ : Optional[int] = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) lowercase__ : Union[str, Any] = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) lowercase__ : Union[str, Any] = lm_model(input_ids=lowercase_ , decoder_input_ids=lowercase_ ) lowercase__ : Optional[int] = (*summary.shape, config.vocab_size) self.assertEqual(outputs["logits"].shape , lowercase_ ) def __UpperCamelCase ( self : List[Any] ) -> str: lowercase__ : List[Any] = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) lowercase__ : int = shift_tokens_right(lowercase_ , 1 , 2 ) lowercase__ : Optional[int] = np.equal(lowercase_ , 1 ).astype(np.floataa ).sum() lowercase__ : Any = np.equal(lowercase_ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowercase_ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class snake_case_ ( __A ,unittest.TestCase ,__A ): __A : List[str] = True __A : List[str] = ( ( FlaxBlenderbotSmallModel, FlaxBlenderbotSmallForConditionalGeneration, ) if is_flax_available() else () ) __A : str = (FlaxBlenderbotSmallForConditionalGeneration,) if is_flax_available() else () def __UpperCamelCase ( self : Any ) -> List[str]: lowercase__ : Dict = FlaxBlenderbotSmallModelTester(self ) def __UpperCamelCase ( self : str ) -> List[str]: lowercase__ , lowercase__ : Any = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowercase_ , lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Any ) -> List[str]: lowercase__ , lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowercase_ , lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Tuple ) -> List[str]: lowercase__ , lowercase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowercase__ : Optional[int] = self._prepare_for_class(lowercase_ , lowercase_ ) lowercase__ : Union[str, Any] = model_class(lowercase_ ) @jax.jit def encode_jitted(lowercase_ : Any , lowercase_ : List[Any]=None , **lowercase_ : Dict ): return model.encode(input_ids=lowercase_ , attention_mask=lowercase_ ) with self.subTest("JIT Enabled" ): lowercase__ : List[str] = encode_jitted(**lowercase_ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): lowercase__ : Any = encode_jitted(**lowercase_ ).to_tuple() self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) for jitted_output, output in zip(lowercase_ , lowercase_ ): self.assertEqual(jitted_output.shape , output.shape ) def __UpperCamelCase ( self : Union[str, Any] ) -> Optional[int]: lowercase__ , lowercase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowercase__ : int = model_class(lowercase_ ) lowercase__ : List[str] = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) lowercase__ : int = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowercase_ : Any , lowercase_ : int , lowercase_ : Optional[int] ): return model.decode( decoder_input_ids=lowercase_ , decoder_attention_mask=lowercase_ , encoder_outputs=lowercase_ , ) with self.subTest("JIT Enabled" ): lowercase__ : Any = decode_jitted(**lowercase_ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): lowercase__ : Tuple = decode_jitted(**lowercase_ ).to_tuple() self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) for jitted_output, output in zip(lowercase_ , lowercase_ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def __UpperCamelCase ( self : Dict ) -> int: for model_class_name in self.all_model_classes: lowercase__ : List[Any] = model_class_name.from_pretrained("facebook/blenderbot_small-90M" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids lowercase__ : Optional[Any] = np.ones((1, 1) ) * model.config.eos_token_id lowercase__ : str = model(lowercase_ ) self.assertIsNotNone(lowercase_ )
87
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] =MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCAmelCase : Union[str, Any] =TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def snake_case ( self , __a , __a , __a ): __lowerCAmelCase = TextaTextGenerationPipeline(model=__a , tokenizer=__a ) return generator, ["Something to write", "Something else"] def snake_case ( self , __a , __a ): __lowerCAmelCase = generator("Something there" ) self.assertEqual(__a , [{"generated_text": ANY(__a )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]["generated_text"].startswith("Something there" ) ) __lowerCAmelCase = generator(["This is great !", "Something else"] , num_return_sequences=2 , do_sample=__a ) self.assertEqual( __a , [ [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], ] , ) __lowerCAmelCase = generator( ["This is great !", "Something else"] , num_return_sequences=2 , batch_size=2 , do_sample=__a ) self.assertEqual( __a , [ [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], ] , ) with self.assertRaises(__a ): generator(4 ) @require_torch def snake_case ( self ): __lowerCAmelCase = pipeline("text2text-generation" , model="patrickvonplaten/t5-tiny-random" , framework="pt" ) # do_sample=False necessary for reproducibility __lowerCAmelCase = generator("Something there" , do_sample=__a ) self.assertEqual(__a , [{"generated_text": ""}] ) __lowerCAmelCase = 3 __lowerCAmelCase = generator( "Something there" , num_return_sequences=__a , num_beams=__a , ) __lowerCAmelCase = [ {"generated_text": "Beide Beide Beide Beide Beide Beide Beide Beide Beide"}, {"generated_text": "Beide Beide Beide Beide Beide Beide Beide Beide"}, {"generated_text": ""}, ] self.assertEqual(__a , __a ) __lowerCAmelCase = generator("This is a test" , do_sample=__a , num_return_sequences=2 , return_tensors=__a ) self.assertEqual( __a , [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ] , ) __lowerCAmelCase = generator.model.config.eos_token_id __lowerCAmelCase = "<pad>" __lowerCAmelCase = generator( ["This is a test", "This is a second test"] , do_sample=__a , num_return_sequences=2 , batch_size=2 , return_tensors=__a , ) self.assertEqual( __a , [ [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ], [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ], ] , ) @require_tf def snake_case ( self ): __lowerCAmelCase = pipeline("text2text-generation" , model="patrickvonplaten/t5-tiny-random" , framework="tf" ) # do_sample=False necessary for reproducibility __lowerCAmelCase = generator("Something there" , do_sample=__a ) self.assertEqual(__a , [{"generated_text": ""}] )
57
0
'''simple docstring''' from functools import lru_cache def UpperCAmelCase__ ( UpperCAmelCase__ ) -> set: A_ = 2 A_ = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(UpperCAmelCase__ ) if n > 1: factors.add(UpperCAmelCase__ ) return factors @lru_cache def UpperCAmelCase__ ( UpperCAmelCase__ ) -> int: return len(unique_prime_factors(UpperCAmelCase__ ) ) def UpperCAmelCase__ ( UpperCAmelCase__ ) -> bool: return len(set(UpperCAmelCase__ ) ) in (0, 1) def UpperCAmelCase__ ( UpperCAmelCase__ ) -> list: A_ = 2 while True: # Increment each value of a generated range A_ = [base + i for i in range(UpperCAmelCase__ )] # Run elements through out unique_prime_factors function # Append our target number to the end. A_ = [upf_len(UpperCAmelCase__ ) for x in group] checker.append(UpperCAmelCase__ ) # If all numbers in the list are equal, return the group variable. if equality(UpperCAmelCase__ ): return group # Increment our base variable by 1 base += 1 def UpperCAmelCase__ ( UpperCAmelCase__ = 4 ) -> int: A_ = run(UpperCAmelCase__ ) return results[0] if len(UpperCAmelCase__ ) else None if __name__ == "__main__": print(solution())
101
'''simple docstring''' import sys __lowerCamelCase = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCAmelCase__ ( UpperCAmelCase__ = N ) -> int: A_ = -sys.maxsize - 1 for i in range(len(UpperCAmelCase__ ) - 12 ): A_ = 1 for j in range(13 ): product *= int(n[i + j] ) if product > largest_product: A_ = product return largest_product if __name__ == "__main__": print(f"""{solution() = }""")
101
1
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Any class a__ : def __init__( self : List[Any] , a : Any ): """simple docstring""" __lowerCamelCase = data __lowerCamelCase = None class a__ : def __init__( self : Any ): """simple docstring""" __lowerCamelCase = None __lowerCamelCase = None def __iter__( self : Any ): """simple docstring""" __lowerCamelCase = self.head while self.head: yield node.data __lowerCamelCase = node.next if node == self.head: break def __len__( self : Any ): """simple docstring""" return sum(1 for _ in self ) def __repr__( self : Union[str, Any] ): """simple docstring""" return "->".join(str(a ) for item in iter(self ) ) def SCREAMING_SNAKE_CASE__ ( self : str , a : Any ): """simple docstring""" self.insert_nth(len(self ) , a ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : Any ): """simple docstring""" self.insert_nth(0 , a ) def SCREAMING_SNAKE_CASE__ ( self : int , a : int , a : Any ): """simple docstring""" if index < 0 or index > len(self ): raise IndexError('''list index out of range.''' ) __lowerCamelCase = Node(a ) if self.head is None: __lowerCamelCase = new_node # first node points itself __lowerCamelCase = __lowerCamelCase = new_node elif index == 0: # insert at head __lowerCamelCase = self.head __lowerCamelCase = __lowerCamelCase = new_node else: __lowerCamelCase = self.head for _ in range(index - 1 ): __lowerCamelCase = temp.next __lowerCamelCase = temp.next __lowerCamelCase = new_node if index == len(self ) - 1: # insert at tail __lowerCamelCase = new_node def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return self.delete_nth(0 ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" return self.delete_nth(len(self ) - 1 ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : int = 0 ): """simple docstring""" if not 0 <= index < len(self ): raise IndexError('''list index out of range.''' ) __lowerCamelCase = self.head if self.head == self.tail: # just one node __lowerCamelCase = __lowerCamelCase = None elif index == 0: # delete head node __lowerCamelCase = self.tail.next.next __lowerCamelCase = self.head.next else: __lowerCamelCase = self.head for _ in range(index - 1 ): __lowerCamelCase = temp.next __lowerCamelCase = temp.next __lowerCamelCase = temp.next.next if index == len(self ) - 1: # delete at tail __lowerCamelCase = temp return delete_node.data def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return len(self ) == 0 def __lowerCAmelCase ( ) -> None: __lowerCamelCase = CircularLinkedList() assert len(UpperCamelCase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(UpperCamelCase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(UpperCamelCase__ ) == i circular_linked_list.insert_nth(UpperCamelCase__ , i + 1 ) assert str(UpperCamelCase__ ) == "->".join(str(UpperCamelCase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(UpperCamelCase__ ) == "->".join(str(UpperCamelCase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(UpperCamelCase__ ) == "->".join(str(UpperCamelCase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(UpperCamelCase__ ) == "->".join(str(UpperCamelCase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(UpperCamelCase__ ) == "->".join(str(UpperCamelCase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
67
'''simple docstring''' def _UpperCAmelCase ( _lowerCamelCase : list[int] , _lowerCamelCase : str ) -> list[int]: _lowerCAmelCase : List[Any] = int(_lowerCamelCase ) # Initialize Result _lowerCAmelCase : Any = [] # Traverse through all denomination for denomination in reversed(_lowerCamelCase ): # Find denominations while int(_lowerCamelCase ) >= int(_lowerCamelCase ): total_value -= int(_lowerCamelCase ) answer.append(_lowerCamelCase ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCamelCase_ = [] UpperCamelCase_ = """0""" if ( input("""Do you want to enter your denominations ? (yY/n): """).strip().lower() == "y" ): UpperCamelCase_ = int(input("""Enter the number of denominations you want to add: """).strip()) for i in range(0, n): denominations.append(int(input(F'Denomination {i}: ').strip())) UpperCamelCase_ = input("""Enter the change you want to make in Indian Currency: """).strip() else: # All denominations of Indian Currency if user does not enter UpperCamelCase_ = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] UpperCamelCase_ = input("""Enter the change you want to make: """).strip() if int(value) == 0 or int(value) < 0: print("""The total value cannot be zero or negative.""") else: print(F'Following is minimal change for {value}: ') UpperCamelCase_ = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=""" """)
309
0
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCAmelCase : Union[str, Any] ={"tokenization_tapex": ["TapexTokenizer"]} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCAmelCase : Dict =_LazyModule(__name__, globals()["__file__"], _import_structure)
123
'''simple docstring''' def UpperCamelCase ( _lowerCamelCase : int ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): A__ = F"Input value of [number={number}] must be an integer" raise TypeError(_lowerCamelCase ) if number < 1: A__ = F"Input value of [number={number}] must be > 0" raise ValueError(_lowerCamelCase ) A__ = 1 for i in range(1 , _lowerCamelCase ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
123
1
from __future__ import annotations import pandas as pd def A_ ( A__ , A__ , A__ ) -> list[int]: a__ : Optional[Any] = [0] * no_of_processes a__ : Tuple = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(A__ ): a__ : Optional[Any] = burst_time[i] a__ : Optional[int] = 0 a__ : int = 0 a__ : int = 9_9999_9999 a__ : Optional[int] = 0 a__ : Tuple = False # Process until all processes are completed while complete != no_of_processes: for j in range(A__ ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: a__ : Dict = remaining_time[j] a__ : int = j a__ : Tuple = True if not check: increment_time += 1 continue remaining_time[short] -= 1 a__ : Tuple = remaining_time[short] if minm == 0: a__ : Optional[Any] = 9_9999_9999 if remaining_time[short] == 0: complete += 1 a__ : Optional[int] = False # Find finish time of current process a__ : Optional[Any] = increment_time + 1 # Calculate waiting time a__ : Optional[Any] = finish_time - arrival_time[short] a__ : Any = finar - burst_time[short] if waiting_time[short] < 0: a__ : int = 0 # Increment time increment_time += 1 return waiting_time def A_ ( A__ , A__ , A__ ) -> list[int]: a__ : List[str] = [0] * no_of_processes for i in range(A__ ): a__ : Dict = burst_time[i] + waiting_time[i] return turn_around_time def A_ ( A__ , A__ , A__ ) -> None: a__ : List[Any] = 0 a__ : Any = 0 for i in range(A__ ): a__ : str = total_waiting_time + waiting_time[i] a__ : Tuple = total_turn_around_time + turn_around_time[i] print(F'Average waiting time = {total_waiting_time / no_of_processes:.5f}' ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") lowercase : str = int(input()) lowercase : List[str] = [0] * no_of_processes lowercase : int = [0] * no_of_processes lowercase : List[Any] = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) lowercase , lowercase : Optional[int] = map(int, input().split()) lowercase : Optional[int] = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowercase : Dict = burst_time lowercase : int = no_of_processes lowercase : Any = waiting_time lowercase : Union[str, Any] = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) lowercase : Any = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
99
'''simple docstring''' import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig UpperCamelCase_ = logging.get_logger(__name__) class a_ : def __init__( self , snake_case_ , snake_case_ ): _lowerCAmelCase : List[str] = question_encoder _lowerCAmelCase : Optional[Any] = generator _lowerCAmelCase : Optional[Any] = self.question_encoder def __UpperCamelCase ( self , snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) _lowerCAmelCase : Any = os.path.join(snake_case_ , """question_encoder_tokenizer""" ) _lowerCAmelCase : Tuple = os.path.join(snake_case_ , """generator_tokenizer""" ) self.question_encoder.save_pretrained(snake_case_ ) self.generator.save_pretrained(snake_case_ ) @classmethod def __UpperCamelCase ( cls , snake_case_ , **snake_case_ ): # dynamically import AutoTokenizer from ..auto.tokenization_auto import AutoTokenizer _lowerCAmelCase : Dict = kwargs.pop("""config""" , snake_case_ ) if config is None: _lowerCAmelCase : List[Any] = RagConfig.from_pretrained(snake_case_ ) _lowerCAmelCase : int = AutoTokenizer.from_pretrained( snake_case_ , config=config.question_encoder , subfolder="""question_encoder_tokenizer""" ) _lowerCAmelCase : Union[str, Any] = AutoTokenizer.from_pretrained( snake_case_ , config=config.generator , subfolder="""generator_tokenizer""" ) return cls(question_encoder=snake_case_ , generator=snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): return self.current_tokenizer(*snake_case_ , **snake_case_ ) def __UpperCamelCase ( self , *snake_case_ , **snake_case_ ): return self.generator.batch_decode(*snake_case_ , **snake_case_ ) def __UpperCamelCase ( self , *snake_case_ , **snake_case_ ): return self.generator.decode(*snake_case_ , **snake_case_ ) def __UpperCamelCase ( self ): _lowerCAmelCase : str = self.question_encoder def __UpperCamelCase ( self ): _lowerCAmelCase : Optional[Any] = self.generator def __UpperCamelCase ( self , snake_case_ , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = "longest" , snake_case_ = None , snake_case_ = True , **snake_case_ , ): warnings.warn( """`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the """ """regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` """ """context manager to prepare your targets. See the documentation of your specific tokenizer for more """ """details""" , snake_case_ , ) if max_length is None: _lowerCAmelCase : Any = self.current_tokenizer.model_max_length _lowerCAmelCase : List[Any] = self( snake_case_ , add_special_tokens=snake_case_ , return_tensors=snake_case_ , max_length=snake_case_ , padding=snake_case_ , truncation=snake_case_ , **snake_case_ , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: _lowerCAmelCase : List[str] = self.current_tokenizer.model_max_length _lowerCAmelCase : List[str] = self( text_target=snake_case_ , add_special_tokens=snake_case_ , return_tensors=snake_case_ , padding=snake_case_ , max_length=snake_case_ , truncation=snake_case_ , **snake_case_ , ) _lowerCAmelCase : Dict = labels["""input_ids"""] return model_inputs
309
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) lowerCAmelCase_ = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['PerceiverFeatureExtractor'] lowerCAmelCase_ = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
365
class _A : # Public class to implement a graph def __init__( self : List[Any] , _A : int , _A : int , _A : list[list[bool]] ) -> None: """simple docstring""" lowercase : Tuple = row lowercase : Union[str, Any] = col lowercase : int = graph def __a ( self : List[Any] , _A : int , _A : int , _A : list[list[bool]] ) -> bool: """simple docstring""" return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def __a ( self : int , _A : int , _A : int , _A : list[list[bool]] ) -> None: """simple docstring""" lowercase : List[str] = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order lowercase : Dict = [-1, 0, 1, -1, 1, -1, 0, 1] lowercase : Dict = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , _A ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , _A ) def __a ( self : List[str] ) -> int: # And finally, count all islands. """simple docstring""" lowercase : List[str] = [[False for j in range(self.COL )] for i in range(self.ROW )] lowercase : Optional[Any] = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(_A , _A , _A ) count += 1 return count
116
0
import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch if is_datasets_available(): from datasets import load_dataset lowercase__ : Tuple = random.Random() def lowerCamelCase__ ( _A , _A=1.0 , _A=None , _A=None ): '''simple docstring''' if rng is None: snake_case_ = global_rng snake_case_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Any , __lowercase : int , __lowercase : Optional[int]=7 , __lowercase : int=4_00 , __lowercase : List[str]=20_00 , __lowercase : Dict=20_48 , __lowercase : int=1_28 , __lowercase : List[Any]=1 , __lowercase : str=5_12 , __lowercase : Any=30 , __lowercase : Dict=4_41_00 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = min_seq_length snake_case_ = max_seq_length snake_case_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) snake_case_ = spectrogram_length snake_case_ = feature_size snake_case_ = num_audio_channels snake_case_ = hop_length snake_case_ = chunk_length snake_case_ = sampling_rate def snake_case__ ( self : List[str] ): """simple docstring""" return { "spectrogram_length": self.spectrogram_length, "feature_size": self.feature_size, "num_audio_channels": self.num_audio_channels, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "sampling_rate": self.sampling_rate, } def snake_case__ ( self : List[Any] , __lowercase : Optional[int]=False , __lowercase : Union[str, Any]=False ): """simple docstring""" def _flatten(__lowercase : str ): return list(itertools.chain(*__lowercase ) ) if equal_length: snake_case_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size snake_case_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: snake_case_ = [np.asarray(__lowercase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class UpperCAmelCase ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = TvltFeatureExtractor def snake_case__ ( self : List[str] ): """simple docstring""" snake_case_ = TvltFeatureExtractionTester(self ) def snake_case__ ( self : Union[str, Any] ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(__lowercase , "spectrogram_length" ) ) self.assertTrue(hasattr(__lowercase , "feature_size" ) ) self.assertTrue(hasattr(__lowercase , "num_audio_channels" ) ) self.assertTrue(hasattr(__lowercase , "hop_length" ) ) self.assertTrue(hasattr(__lowercase , "chunk_length" ) ) self.assertTrue(hasattr(__lowercase , "sampling_rate" ) ) def snake_case__ ( self : Optional[int] ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = feat_extract_first.save_pretrained(__lowercase )[0] check_json_file_has_correct_format(__lowercase ) snake_case_ = self.feature_extraction_class.from_pretrained(__lowercase ) snake_case_ = feat_extract_first.to_dict() snake_case_ = feat_extract_second.to_dict() snake_case_ = dict_first.pop("mel_filters" ) snake_case_ = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(__lowercase , __lowercase ) ) self.assertEqual(__lowercase , __lowercase ) def snake_case__ ( self : Dict ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = os.path.join(__lowercase , "feat_extract.json" ) feat_extract_first.to_json_file(__lowercase ) snake_case_ = self.feature_extraction_class.from_json_file(__lowercase ) snake_case_ = feat_extract_first.to_dict() snake_case_ = feat_extract_second.to_dict() snake_case_ = dict_first.pop("mel_filters" ) snake_case_ = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(__lowercase , __lowercase ) ) self.assertEqual(__lowercase , __lowercase ) def snake_case__ ( self : Dict ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) # create three inputs of length 800, 1000, and 1200 snake_case_ = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] snake_case_ = [np.asarray(__lowercase ) for speech_input in speech_inputs] # Test not batched input snake_case_ = feature_extractor(np_speech_inputs[0] , return_tensors="np" , sampling_rate=4_41_00 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test batched snake_case_ = feature_extractor(__lowercase , return_tensors="np" , sampling_rate=4_41_00 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test audio masking snake_case_ = feature_extractor( __lowercase , return_tensors="np" , sampling_rate=4_41_00 , mask_audio=__lowercase ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test 2-D numpy arrays are batched. snake_case_ = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] snake_case_ = np.asarray(__lowercase ) snake_case_ = feature_extractor(__lowercase , return_tensors="np" , sampling_rate=4_41_00 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) def snake_case__ ( self : str , __lowercase : Tuple ): """simple docstring""" snake_case_ = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech snake_case_ = ds.sort("id" ).select(range(__lowercase ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def snake_case__ ( self : Union[str, Any] ): """simple docstring""" snake_case_ = self._load_datasamples(1 ) snake_case_ = TvltFeatureExtractor() snake_case_ = feature_extractor(__lowercase , return_tensors="pt" ).audio_values self.assertEquals(audio_values.shape , (1, 1, 1_92, 1_28) ) snake_case_ = torch.tensor([[-0.3032, -0.2708], [-0.4434, -0.4007]] ) self.assertTrue(torch.allclose(audio_values[0, 0, :2, :2] , __lowercase , atol=1E-4 ) )
187
def lowerCamelCase__ ( ): '''simple docstring''' snake_case_ = [] snake_case_ = 1 while len(_A ) < 1E6: constant.append(str(_A ) ) i += 1 snake_case_ = "".join(_A ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[99999] ) * int(constant[999999] ) ) if __name__ == "__main__": print(solution())
187
1
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Dict = logging.get_logger(__name__) __magic_name__ : int = { "snap-research/efficientformer-l1-300": ( "https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json" ), } class SCREAMING_SNAKE_CASE_ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowercase : str = '''efficientformer''' def __init__( self , lowerCAmelCase__ = [3, 2, 6, 4] , lowerCAmelCase__ = [4_8, 9_6, 2_2_4, 4_4_8] , lowerCAmelCase__ = [True, True, True, True] , lowerCAmelCase__ = 4_4_8 , lowerCAmelCase__ = 3_2 , lowerCAmelCase__ = 4 , lowerCAmelCase__ = 7 , lowerCAmelCase__ = 5 , lowerCAmelCase__ = 8 , lowerCAmelCase__ = 4 , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1_6 , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = 1E-5 , lowerCAmelCase__ = "gelu" , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 1E-12 , lowerCAmelCase__ = 2_2_4 , lowerCAmelCase__ = 1E-05 , **lowerCAmelCase__ , ): super().__init__(**lowerCAmelCase__) __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = hidden_sizes __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = patch_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = depths __SCREAMING_SNAKE_CASE = mlp_expansion_ratio __SCREAMING_SNAKE_CASE = downsamples __SCREAMING_SNAKE_CASE = dim __SCREAMING_SNAKE_CASE = key_dim __SCREAMING_SNAKE_CASE = attention_ratio __SCREAMING_SNAKE_CASE = resolution __SCREAMING_SNAKE_CASE = pool_size __SCREAMING_SNAKE_CASE = downsample_patch_size __SCREAMING_SNAKE_CASE = downsample_stride __SCREAMING_SNAKE_CASE = downsample_pad __SCREAMING_SNAKE_CASE = drop_path_rate __SCREAMING_SNAKE_CASE = num_metaad_blocks __SCREAMING_SNAKE_CASE = distillation __SCREAMING_SNAKE_CASE = use_layer_scale __SCREAMING_SNAKE_CASE = layer_scale_init_value __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = batch_norm_eps
351
"""simple docstring""" import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=9_9 , lowerCAmelCase__=3_2 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=3_7 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1_2_8 , lowerCAmelCase__=3_2 , lowerCAmelCase__=1_6 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=None , ): __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_input_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = num_choices __SCREAMING_SNAKE_CASE = scope def snake_case_ ( self): __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __SCREAMING_SNAKE_CASE = None if self.use_input_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length]) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices) __SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case_ ( self): return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , ) def snake_case_ ( self): ( ( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) , ) = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = NezhaModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = NezhaModel(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = NezhaForMaskedLM(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = NezhaForNextSentencePrediction(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = NezhaForPreTraining(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ , next_sentence_label=lowerCAmelCase__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = NezhaForQuestionAnswering(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , start_positions=lowerCAmelCase__ , end_positions=lowerCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = NezhaForSequenceClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = NezhaForTokenClassification(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_choices __SCREAMING_SNAKE_CASE = NezhaForMultipleChoice(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) , ) = config_and_inputs __SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( __a , __a , __a , unittest.TestCase ): """simple docstring""" __lowercase : int = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) __lowercase : Optional[Any] = ( { '''feature-extraction''': NezhaModel, '''fill-mask''': NezhaForMaskedLM, '''question-answering''': NezhaForQuestionAnswering, '''text-classification''': NezhaForSequenceClassification, '''token-classification''': NezhaForTokenClassification, '''zero-shot''': NezhaForSequenceClassification, } if is_torch_available() else {} ) __lowercase : List[Any] = True def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False): __SCREAMING_SNAKE_CASE = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) if return_labels: if model_class in get_values(lowerCAmelCase__): __SCREAMING_SNAKE_CASE = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCAmelCase__) return inputs_dict def snake_case_ ( self): __SCREAMING_SNAKE_CASE = NezhaModelTester(self) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=3_7) def snake_case_ ( self): self.config_tester.run_common_tests() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCAmelCase__) def snake_case_ ( self): # This regression test was failing with PyTorch < 1.3 ( ( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() __SCREAMING_SNAKE_CASE = None self.model_tester.create_and_check_model_as_decoder( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCAmelCase__) @slow def snake_case_ ( self): for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = NezhaModel.from_pretrained(lowerCAmelCase__) self.assertIsNotNone(lowerCAmelCase__) @slow @require_torch_gpu def snake_case_ ( self): __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = model_class(config=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.jit.trace( lowerCAmelCase__ , (inputs_dict["""input_ids"""].to("""cpu"""), inputs_dict["""attention_mask"""].to("""cpu"""))) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , """bert.pt""")) __SCREAMING_SNAKE_CASE = torch.jit.load(os.path.join(lowerCAmelCase__ , """bert.pt""") , map_location=lowerCAmelCase__) loaded(inputs_dict["""input_ids"""].to(lowerCAmelCase__) , inputs_dict["""attention_mask"""].to(lowerCAmelCase__)) @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = NezhaModel.from_pretrained("""sijunhe/nezha-cn-base""") __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]]) __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 1, 1, 1, 1]]) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__)[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 6, 7_6_8)) self.assertEqual(output.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor([[[0.06_85, 0.24_41, 0.11_02], [0.06_00, 0.19_06, 0.13_49], [0.02_21, 0.08_19, 0.05_86]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase__ , atol=1E-4)) @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = NezhaForMaskedLM.from_pretrained("""sijunhe/nezha-cn-base""") __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]]) __SCREAMING_SNAKE_CASE = torch.tensor([[1, 1, 1, 1, 1, 1]]) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__)[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 6, 2_1_1_2_8)) self.assertEqual(output.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor( [[-2.79_39, -1.79_02, -2.21_89], [-2.85_85, -1.89_08, -2.37_23], [-2.64_99, -1.77_50, -2.25_58]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase__ , atol=1E-4))
255
0
from ...configuration_utils import PretrainedConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : int = { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json", # See all REALM models at https://huggingface.co/models?filter=realm } class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = '''realm''' def __init__( self : int , __magic_name__ : int=30_522 , __magic_name__ : Optional[int]=768 , __magic_name__ : Dict=128 , __magic_name__ : int=12 , __magic_name__ : Optional[Any]=12 , __magic_name__ : Optional[Any]=8 , __magic_name__ : str=3_072 , __magic_name__ : Union[str, Any]="gelu_new" , __magic_name__ : int=0.1 , __magic_name__ : List[Any]=0.1 , __magic_name__ : str=512 , __magic_name__ : str=2 , __magic_name__ : Tuple=0.02 , __magic_name__ : Optional[int]=1e-12 , __magic_name__ : List[Any]=256 , __magic_name__ : Any=10 , __magic_name__ : Optional[Any]=1e-3 , __magic_name__ : str=5 , __magic_name__ : List[str]=320 , __magic_name__ : Optional[int]=13_353_718 , __magic_name__ : Dict=5_000 , __magic_name__ : Optional[int]=1 , __magic_name__ : Union[str, Any]=0 , __magic_name__ : Tuple=2 , **__magic_name__ : Dict , ) -> Union[str, Any]: super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) # Common config SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = max_position_embeddings SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = retriever_proj_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = num_candidates SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = type_vocab_size SCREAMING_SNAKE_CASE_ = layer_norm_eps # Reader config SCREAMING_SNAKE_CASE_ = span_hidden_size SCREAMING_SNAKE_CASE_ = max_span_width SCREAMING_SNAKE_CASE_ = reader_layer_norm_eps SCREAMING_SNAKE_CASE_ = reader_beam_size SCREAMING_SNAKE_CASE_ = reader_seq_len # Retrieval config SCREAMING_SNAKE_CASE_ = num_block_records SCREAMING_SNAKE_CASE_ = searcher_beam_size
118
import math from datetime import datetime, timedelta def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = year % 1_9 SCREAMING_SNAKE_CASE_ = year % 4 SCREAMING_SNAKE_CASE_ = year % 7 SCREAMING_SNAKE_CASE_ = math.floor(year / 1_0_0 ) SCREAMING_SNAKE_CASE_ = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) SCREAMING_SNAKE_CASE_ = leap_day_inhibits / 4 SCREAMING_SNAKE_CASE_ = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 SCREAMING_SNAKE_CASE_ = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 SCREAMING_SNAKE_CASE_ = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon SCREAMING_SNAKE_CASE_ = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(__UpperCamelCase , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(__UpperCamelCase , 4 , 1_8 ) else: return datetime(__UpperCamelCase , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (19_94, 20_00, 20_10, 20_21, 20_23): A : Dict = "will be" if year > datetime.now().year else "was" print(f"Easter in {year} {tense} {gauss_easter(year)}")
118
1
"""simple docstring""" from torch import nn class __UpperCAmelCase( nn.Module ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): '''simple docstring''' super().__init__() lowercase__ : Optional[Any]= class_size lowercase__ : int= embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) lowercase__ : Dict= nn.Linear(snake_case__ , snake_case__ ) def UpperCAmelCase_ ( self , snake_case__ ): '''simple docstring''' # hidden_state = nn.functional.relu(self.mlp1(hidden_state)) # hidden_state = self.mlp2(hidden_state) lowercase__ : str= self.mlp(snake_case__ ) return logits
150
"""simple docstring""" from __future__ import annotations def lowercase__(A ) ->int: """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(A ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(A ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
150
1
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = {"""vocab_file""": """spiece.model"""} _SCREAMING_SNAKE_CASE = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } _SCREAMING_SNAKE_CASE = { """AI-Sweden/gpt-sw3-126m""": 2_0_4_8, """AI-Sweden/gpt-sw3-350m""": 2_0_4_8, """AI-Sweden/gpt-sw3-1.6b""": 2_0_4_8, """AI-Sweden/gpt-sw3-6.7b""": 2_0_4_8, """AI-Sweden/gpt-sw3-20b""": 2_0_4_8, } class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["""input_ids""", """attention_mask"""] def __init__( self : str , lowerCamelCase_ : int , lowerCamelCase_ : Optional[int]=False , lowerCamelCase_ : Union[str, Any]=False , lowerCamelCase_ : Tuple=False , lowerCamelCase_ : Optional[int]=None , lowerCamelCase_ : int=None , lowerCamelCase_ : Optional[Any]=None , lowerCamelCase_ : List[Any]=None , lowerCamelCase_ : Optional[Dict[str, Any]] = None , **lowerCamelCase_ : str , ): """simple docstring""" UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs UpperCamelCase = kwargs.get("""name_or_path""" ) if name_or_path is None: logger.warning( """name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,""" """ you are testing the model, this can safely be ignored""" ) UpperCamelCase = """None""" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing UpperCamelCase = """<|endoftext|>""" if eos_token is None else eos_token UpperCamelCase = """<unk>""" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: UpperCamelCase = unk_token if pad_token is None else pad_token UpperCamelCase = eos_token if bos_token is None else bos_token else: UpperCamelCase = """<pad>""" if pad_token is None else pad_token UpperCamelCase = """<s>""" if bos_token is None else bos_token super().__init__( do_lower_case=lowerCamelCase_ , remove_space=lowerCamelCase_ , keep_accents=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) UpperCamelCase = do_lower_case UpperCamelCase = remove_space UpperCamelCase = keep_accents UpperCamelCase = vocab_file UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase_ ) # Used for whitespace normalization in input texts # fmt : off UpperCamelCase = {""" """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """""", """„"""} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing UpperCamelCase = re.compile( f"""[{"".join(map(lowerCamelCase_ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = self.__dict__.copy() UpperCamelCase = None return state def __setstate__( self : Any , lowerCamelCase_ : List[str] ): """simple docstring""" UpperCamelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCamelCase = {} UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" return len(self.sp_model ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : str ): """simple docstring""" UpperCamelCase = self.non_printing_characters_re.sub("""""" , lowerCamelCase_ ) # Normalize whitespaces UpperCamelCase = """""".join([char if char not in self.whitespaces else """ """ for char in text] ) # NFC Unicode normalization UpperCamelCase = unicodedata.normalize("""NFC""" , lowerCamelCase_ ) return text def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : str , **lowerCamelCase_ : Tuple ): """simple docstring""" UpperCamelCase = self.preprocess_text(lowerCamelCase_ ) return self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_ ) def lowerCamelCase_ ( self : str , lowerCamelCase_ : str ): """simple docstring""" return self.sp_model.PieceToId(lowerCamelCase_ ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : int ): """simple docstring""" return self.sp_model.IdToPiece(lowerCamelCase_ ) @staticmethod def lowerCamelCase_ ( lowerCamelCase_ : str ): """simple docstring""" return out_string def lowerCamelCase_ ( self : List[str] , lowerCamelCase_ : List[str] ): """simple docstring""" UpperCamelCase = [] UpperCamelCase = """""" UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCamelCase_ ) + token UpperCamelCase = True UpperCamelCase = [] else: current_sub_tokens.append(lowerCamelCase_ ) UpperCamelCase = False out_string += self.sp_model.decode(lowerCamelCase_ ) return out_string def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = {self.convert_ids_to_tokens(lowerCamelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase_ ( self : int , lowerCamelCase_ : str , lowerCamelCase_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase_ , """wb""" ) as fi: UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_ ) return (out_vocab_file,) def lowerCamelCase_ ( self : List[Any] , lowerCamelCase_ : Union[str, List[str]] , lowerCamelCase_ : Union[str, bool] = False ): """simple docstring""" if isinstance(lowerCamelCase_ , lowerCamelCase_ ): UpperCamelCase = self.preprocess_text(lowerCamelCase_ ) UpperCamelCase = self.sp_model.encode(lowerCamelCase_ ) else: UpperCamelCase = [self.preprocess_text(lowerCamelCase_ ) for t in text] UpperCamelCase = self.sp_model.encode(lowerCamelCase_ ) if return_tensors is True or return_tensors == "pt": UpperCamelCase = torch.tensor(lowerCamelCase_ ) return token_ids def lowerCamelCase_ ( self : int , lowerCamelCase_ : Union[int, List[int]] ): """simple docstring""" return self.sp_model.decode(lowerCamelCase_ ) def lowerCamelCase_ ( self : List[Any] , lowerCamelCase_ : "Conversation" ): """simple docstring""" UpperCamelCase = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] UpperCamelCase = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(lowerCamelCase_ ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=lowerCamelCase_ )
343
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} # See all LED models at https://huggingface.co/models?filter=LED _SCREAMING_SNAKE_CASE = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } _SCREAMING_SNAKE_CASE = { """allenai/led-base-16384""": 1_6_3_8_4, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def lowercase( ) -> List[str]: '''simple docstring''' UpperCamelCase = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) UpperCamelCase = bs[:] UpperCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCamelCase_ ) cs.append(2**8 + n ) n += 1 UpperCamelCase = [chr(UpperCamelCase_ ) for n in cs] return dict(zip(UpperCamelCase_ , UpperCamelCase_ ) ) def lowercase( UpperCamelCase_ ) -> List[str]: '''simple docstring''' UpperCamelCase = set() UpperCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCamelCase = char return pairs class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["""input_ids""", """attention_mask"""] def __init__( self : str , lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Optional[int] , lowerCamelCase_ : str="replace" , lowerCamelCase_ : Any="<s>" , lowerCamelCase_ : List[Any]="</s>" , lowerCamelCase_ : List[Any]="</s>" , lowerCamelCase_ : str="<s>" , lowerCamelCase_ : str="<unk>" , lowerCamelCase_ : int="<pad>" , lowerCamelCase_ : List[str]="<mask>" , lowerCamelCase_ : str=False , **lowerCamelCase_ : str , ): """simple docstring""" UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else bos_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else eos_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else sep_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else cls_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else unk_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else mask_token super().__init__( errors=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , add_prefix_space=lowerCamelCase_ , **lowerCamelCase_ , ) with open(lowerCamelCase_ , encoding="""utf-8""" ) as vocab_handle: UpperCamelCase = json.load(lowerCamelCase_ ) UpperCamelCase = {v: k for k, v in self.encoder.items()} UpperCamelCase = errors # how to handle errors in decoding UpperCamelCase = bytes_to_unicode() UpperCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase_ , encoding="""utf-8""" ) as merges_handle: UpperCamelCase = merges_handle.read().split("""\n""" )[1:-1] UpperCamelCase = [tuple(merge.split() ) for merge in bpe_merges] UpperCamelCase = dict(zip(lowerCamelCase_ , range(len(lowerCamelCase_ ) ) ) ) UpperCamelCase = {} UpperCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCamelCase = re.compile(R"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def lowerCamelCase_ ( self : str ): """simple docstring""" return len(self.encoder ) def lowerCamelCase_ ( self : str ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase_ ( self : Dict , lowerCamelCase_ : Dict ): """simple docstring""" if token in self.cache: return self.cache[token] UpperCamelCase = tuple(lowerCamelCase_ ) UpperCamelCase = get_pairs(lowerCamelCase_ ) if not pairs: return token while True: UpperCamelCase = min(lowerCamelCase_ , key=lambda lowerCamelCase_ : self.bpe_ranks.get(lowerCamelCase_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break UpperCamelCase , UpperCamelCase = bigram UpperCamelCase = [] UpperCamelCase = 0 while i < len(lowerCamelCase_ ): try: UpperCamelCase = word.index(lowerCamelCase_ , lowerCamelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCamelCase = j if word[i] == first and i < len(lowerCamelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCamelCase = tuple(lowerCamelCase_ ) UpperCamelCase = new_word if len(lowerCamelCase_ ) == 1: break else: UpperCamelCase = get_pairs(lowerCamelCase_ ) UpperCamelCase = """ """.join(lowerCamelCase_ ) UpperCamelCase = word return word def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : Tuple ): """simple docstring""" UpperCamelCase = [] for token in re.findall(self.pat , lowerCamelCase_ ): UpperCamelCase = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase_ ).split(""" """ ) ) return bpe_tokens def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : str ): """simple docstring""" return self.encoder.get(lowerCamelCase_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : Any ): """simple docstring""" return self.decoder.get(lowerCamelCase_ ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : str ): """simple docstring""" UpperCamelCase = """""".join(lowerCamelCase_ ) UpperCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def lowerCamelCase_ ( self : int , lowerCamelCase_ : str , lowerCamelCase_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCamelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase_ , ensure_ascii=lowerCamelCase_ ) + """\n""" ) UpperCamelCase = 0 with open(lowerCamelCase_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase_ : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" """ Please check that the tokenizer is not corrupted!""" ) UpperCamelCase = token_index writer.write(""" """.join(lowerCamelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file def lowerCamelCase_ ( self : str , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCamelCase = [self.cls_token_id] UpperCamelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCamelCase_ ( self : Tuple , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None , lowerCamelCase_ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase_ )) + [1] return [1] + ([0] * len(lowerCamelCase_ )) + [1, 1] + ([0] * len(lowerCamelCase_ )) + [1] def lowerCamelCase_ ( self : List[str] , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None ): """simple docstring""" UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase_ ( self : str , lowerCamelCase_ : Tuple , lowerCamelCase_ : Optional[int]=False , **lowerCamelCase_ : Any ): """simple docstring""" UpperCamelCase = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase_ ) > 0 and not text[0].isspace()): UpperCamelCase = """ """ + text return (text, kwargs) def lowerCamelCase_ ( self : Optional[int] , lowerCamelCase_ : Union[Dict[str, EncodedInput], BatchEncoding] , lowerCamelCase_ : Optional[int] = None , lowerCamelCase_ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , lowerCamelCase_ : Optional[int] = None , lowerCamelCase_ : Optional[bool] = None , ): """simple docstring""" UpperCamelCase = super()._pad( encoded_inputs=lowerCamelCase_ , max_length=lowerCamelCase_ , padding_strategy=lowerCamelCase_ , pad_to_multiple_of=lowerCamelCase_ , return_attention_mask=lowerCamelCase_ , ) # Load from model defaults if return_attention_mask is None: UpperCamelCase = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: UpperCamelCase = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. UpperCamelCase = len(encoded_inputs["""global_attention_mask"""] ) != len(lowerCamelCase_ ) if needs_to_be_padded: UpperCamelCase = len(lowerCamelCase_ ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` UpperCamelCase = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": UpperCamelCase = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
343
1
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = 42 class _UpperCAmelCase ( snake_case_ , snake_case_ ): """simple docstring""" @register_to_config def __init__( self : int , __UpperCAmelCase : int = 3 , __UpperCAmelCase : int = 3 , __UpperCAmelCase : Tuple[str] = ("DownEncoderBlock2D",) , __UpperCAmelCase : Tuple[str] = ("UpDecoderBlock2D",) , __UpperCAmelCase : Tuple[int] = (64,) , __UpperCAmelCase : int = 1 , __UpperCAmelCase : str = "silu" , __UpperCAmelCase : int = 3 , __UpperCAmelCase : int = 32 , __UpperCAmelCase : int = 256 , __UpperCAmelCase : int = 32 , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : float = 0.18215 , __UpperCAmelCase : str = "group" , ): '''simple docstring''' super().__init__() # pass init params to Encoder _A = Encoder( in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , down_block_types=__UpperCAmelCase , block_out_channels=__UpperCAmelCase , layers_per_block=__UpperCAmelCase , act_fn=__UpperCAmelCase , norm_num_groups=__UpperCAmelCase , double_z=__UpperCAmelCase , ) _A = vq_embed_dim if vq_embed_dim is not None else latent_channels _A = nn.Convad(__UpperCAmelCase , __UpperCAmelCase , 1 ) _A = VectorQuantizer(__UpperCAmelCase , __UpperCAmelCase , beta=0.25 , remap=__UpperCAmelCase , sane_index_shape=__UpperCAmelCase ) _A = nn.Convad(__UpperCAmelCase , __UpperCAmelCase , 1 ) # pass init params to Decoder _A = Decoder( in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , up_block_types=__UpperCAmelCase , block_out_channels=__UpperCAmelCase , layers_per_block=__UpperCAmelCase , act_fn=__UpperCAmelCase , norm_num_groups=__UpperCAmelCase , norm_type=__UpperCAmelCase , ) @apply_forward_hook def lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : bool = True ): '''simple docstring''' _A = self.encoder(__UpperCAmelCase ) _A = self.quant_conv(__UpperCAmelCase ) if not return_dict: return (h,) return VQEncoderOutput(latents=__UpperCAmelCase ) @apply_forward_hook def lowerCAmelCase ( self : str , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : bool = False , __UpperCAmelCase : bool = True ): '''simple docstring''' if not force_not_quantize: _A , _A , _A = self.quantize(__UpperCAmelCase ) else: _A = h _A = self.post_quant_conv(__UpperCAmelCase ) _A = self.decoder(__UpperCAmelCase , quant if self.config.norm_type == "spatial" else None ) if not return_dict: return (dec,) return DecoderOutput(sample=__UpperCAmelCase ) def lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : bool = True ): '''simple docstring''' _A = sample _A = self.encode(__UpperCAmelCase ).latents _A = self.decode(__UpperCAmelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=__UpperCAmelCase )
174
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ = {'''configuration_mbart''': ['''MBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MBartConfig''', '''MBartOnnxConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['''MBartTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['''MBartTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''MBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MBartForCausalLM''', '''MBartForConditionalGeneration''', '''MBartForQuestionAnswering''', '''MBartForSequenceClassification''', '''MBartModel''', '''MBartPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''TFMBartForConditionalGeneration''', '''TFMBartModel''', '''TFMBartPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''FlaxMBartForConditionalGeneration''', '''FlaxMBartForQuestionAnswering''', '''FlaxMBartForSequenceClassification''', '''FlaxMBartModel''', '''FlaxMBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
174
1
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer lowerCamelCase : Optional[int] = ['bert-base-uncased', 'bert-base-cased'] lowerCamelCase : Union[str, Any] = 'hf-internal-testing/tiny-bert-tf-only' if is_tf_available(): class __lowerCAmelCase (tf.keras.Model ): '''simple docstring''' def __init__(self : List[str] , UpperCamelCase : Optional[Any] ): '''simple docstring''' super().__init__() lowercase__ = tokenizer lowercase__ = AutoConfig.from_pretrained(UpperCamelCase ) lowercase__ = TFAutoModel.from_config(UpperCamelCase ) def UpperCamelCase__ (self : List[str] , UpperCamelCase : List[str] ): '''simple docstring''' lowercase__ = self.tokenizer(UpperCamelCase ) lowercase__ = self.bert(**UpperCamelCase ) return out["pooler_output"] @require_tf @require_tensorflow_text class __lowerCAmelCase (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ (self : Tuple ): '''simple docstring''' super().setUp() lowercase__ = [ BertTokenizer.from_pretrained(UpperCamelCase ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false lowercase__ = [TFBertTokenizer.from_pretrained(UpperCamelCase ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(UpperCamelCase , use_fast_bert_tokenizer=UpperCamelCase ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) lowercase__ = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] lowercase__ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def UpperCamelCase__ (self : Union[str, Any] ): '''simple docstring''' for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): lowercase__ = tokenizer(UpperCamelCase , return_tensors='''tf''' , padding='''longest''' ) lowercase__ = tf_tokenizer(UpperCamelCase ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def UpperCamelCase__ (self : int ): '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: lowercase__ = tf_tokenizer(self.paired_sentences ) lowercase__ = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def UpperCamelCase__ (self : Any ): '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: lowercase__ = tf.function(UpperCamelCase ) for test_inputs in (self.test_sentences, self.paired_sentences): lowercase__ = tf.constant(UpperCamelCase ) lowercase__ = compiled_tokenizer(UpperCamelCase ) lowercase__ = tf_tokenizer(UpperCamelCase ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def UpperCamelCase__ (self : Tuple ): '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: lowercase__ = ModelToSave(tokenizer=UpperCamelCase ) lowercase__ = tf.convert_to_tensor(self.test_sentences ) lowercase__ = model(UpperCamelCase ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: lowercase__ = Path(UpperCamelCase ) / '''saved.model''' model.save(UpperCamelCase ) lowercase__ = tf.keras.models.load_model(UpperCamelCase ) lowercase__ = loaded_model(UpperCamelCase ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
2
'''simple docstring''' from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : List[Any] = logging.get_logger(__name__) lowerCamelCase : str = { "huggingface/time-series-transformer-tourism-monthly": ( "https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json" ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class A__ ( A__ ): A__ = 'time_series_transformer' A__ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self : Optional[int] , _a : Optional[int] = None , _a : Optional[int] = None , _a : str = "student_t" , _a : str = "nll" , _a : int = 1 , _a : List[int] = [1, 2, 3, 4, 5, 6, 7] , _a : Optional[Union[str, bool]] = "mean" , _a : int = 0 , _a : int = 0 , _a : int = 0 , _a : int = 0 , _a : Optional[List[int]] = None , _a : Optional[List[int]] = None , _a : int = 32 , _a : int = 32 , _a : int = 2 , _a : int = 2 , _a : int = 2 , _a : int = 2 , _a : bool = True , _a : str = "gelu" , _a : int = 64 , _a : float = 0.1 , _a : float = 0.1 , _a : float = 0.1 , _a : float = 0.1 , _a : float = 0.1 , _a : int = 100 , _a : float = 0.02 , _a : Union[str, Any]=True , **_a : Optional[Any] , ) -> Optional[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =prediction_length _SCREAMING_SNAKE_CASE =context_length or prediction_length _SCREAMING_SNAKE_CASE =distribution_output _SCREAMING_SNAKE_CASE =loss _SCREAMING_SNAKE_CASE =input_size _SCREAMING_SNAKE_CASE =num_time_features _SCREAMING_SNAKE_CASE =lags_sequence _SCREAMING_SNAKE_CASE =scaling _SCREAMING_SNAKE_CASE =num_dynamic_real_features _SCREAMING_SNAKE_CASE =num_static_real_features _SCREAMING_SNAKE_CASE =num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(_a ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) _SCREAMING_SNAKE_CASE =cardinality else: _SCREAMING_SNAKE_CASE =[0] if embedding_dimension and num_static_categorical_features > 0: if len(_a ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) _SCREAMING_SNAKE_CASE =embedding_dimension else: _SCREAMING_SNAKE_CASE =[min(50 , (cat + 1) // 2 ) for cat in self.cardinality] _SCREAMING_SNAKE_CASE =num_parallel_samples # Transformer architecture configuration _SCREAMING_SNAKE_CASE =input_size * len(_a ) + self._number_of_features _SCREAMING_SNAKE_CASE =d_model _SCREAMING_SNAKE_CASE =encoder_attention_heads _SCREAMING_SNAKE_CASE =decoder_attention_heads _SCREAMING_SNAKE_CASE =encoder_ffn_dim _SCREAMING_SNAKE_CASE =decoder_ffn_dim _SCREAMING_SNAKE_CASE =encoder_layers _SCREAMING_SNAKE_CASE =decoder_layers _SCREAMING_SNAKE_CASE =dropout _SCREAMING_SNAKE_CASE =attention_dropout _SCREAMING_SNAKE_CASE =activation_dropout _SCREAMING_SNAKE_CASE =encoder_layerdrop _SCREAMING_SNAKE_CASE =decoder_layerdrop _SCREAMING_SNAKE_CASE =activation_function _SCREAMING_SNAKE_CASE =init_std _SCREAMING_SNAKE_CASE =use_cache super().__init__(is_encoder_decoder=_a , **_a ) @property def A ( self : List[Any] ) -> int: '''simple docstring''' return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
47
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { '''facebook/s2t-wav2vec2-large-en-de''': ( '''https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json''' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class _snake_case ( a__ ): snake_case__ = "speech_to_text_2" snake_case__ = ["past_key_values"] snake_case__ = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : List[Any] , UpperCAmelCase : Dict=10000 , UpperCAmelCase : Optional[Any]=6 , UpperCAmelCase : Any=2048 , UpperCAmelCase : Optional[Any]=4 , UpperCAmelCase : List[str]=0.0 , UpperCAmelCase : str=True , UpperCAmelCase : str="relu" , UpperCAmelCase : Any=256 , UpperCAmelCase : Optional[Any]=0.1 , UpperCAmelCase : Optional[Any]=0.0 , UpperCAmelCase : List[Any]=0.0 , UpperCAmelCase : Tuple=0.0_2 , UpperCAmelCase : Dict=2 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=1 , UpperCAmelCase : int=0 , UpperCAmelCase : Optional[int]=2 , UpperCAmelCase : Union[str, Any]=1024 , **UpperCAmelCase : int , ): __lowerCamelCase : int = vocab_size __lowerCamelCase : List[Any] = d_model __lowerCamelCase : Union[str, Any] = decoder_ffn_dim __lowerCamelCase : Tuple = decoder_layers __lowerCamelCase : Dict = decoder_attention_heads __lowerCamelCase : Any = dropout __lowerCamelCase : List[Any] = attention_dropout __lowerCamelCase : Any = activation_dropout __lowerCamelCase : Dict = activation_function __lowerCamelCase : Optional[Any] = init_std __lowerCamelCase : Optional[Any] = decoder_layerdrop __lowerCamelCase : int = use_cache __lowerCamelCase : List[Any] = decoder_layers __lowerCamelCase : str = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCamelCase : Any = max_target_positions super().__init__( pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , decoder_start_token_id=UpperCAmelCase , **UpperCAmelCase , )
64
"""simple docstring""" import requests __A = '''YOUR API KEY''' def lowercase_ ( _lowerCamelCase: str , _lowerCamelCase: str = giphy_api_key ) -> list: '''simple docstring''' __lowerCamelCase : Dict = "+".join(query.split() ) __lowerCamelCase : Optional[int] = F"""https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}""" __lowerCamelCase : Optional[Any] = requests.get(_lowerCamelCase ).json()["data"] return [gif["url"] for gif in gifs] if __name__ == "__main__": print('''\n'''.join(get_gifs('''space ship''')))
64
1
'''simple docstring''' from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class UpperCAmelCase_ ( __lowercase ): def __lt__( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> List[Any]: return self[-1] < other[-1] def __eq__( self : str , UpperCAmelCase__ : List[str] ) -> Tuple: return self[-1] == other[-1] def a_ ( lowerCamelCase : list ): lowerCAmelCase = [] # sort into stacks for element in collection: lowerCAmelCase = Stack([element] ) lowerCAmelCase = bisect_left(lowerCamelCase , lowerCamelCase ) if i != len(lowerCamelCase ): stacks[i].append(lowerCamelCase ) else: stacks.append(lowerCamelCase ) # use a heap-based merge to merge stack efficiently lowerCAmelCase = merge(*(reversed(lowerCamelCase ) for stack in stacks) ) return collection if __name__ == "__main__": __snake_case =input("""Enter numbers separated by a comma:\n""").strip() __snake_case =[int(item) for item in user_input.split(""",""")] print(patience_sort(unsorted))
4
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() __snake_case =logging.get_logger("""transformers.models.encodec""") __snake_case ={ """quantizer.vq.layers.*._codebook.inited""": """quantizer.layers.*.codebook.inited""", """quantizer.vq.layers.*._codebook.cluster_size""": """quantizer.layers.*.codebook.cluster_size""", """quantizer.vq.layers.*._codebook.embed""": """quantizer.layers.*.codebook.embed""", """quantizer.vq.layers.*._codebook.embed_avg""": """quantizer.layers.*.codebook.embed_avg""", } __snake_case ={ """encoder.model.0.conv.conv""": """encoder.layers.0.conv""", """encoder.model.1.block.1.conv.conv""": """encoder.layers.1.block.1.conv""", """encoder.model.1.block.3.conv.conv""": """encoder.layers.1.block.3.conv""", """encoder.model.1.shortcut.conv.conv""": """encoder.layers.1.shortcut.conv""", """encoder.model.3.conv.conv""": """encoder.layers.3.conv""", """encoder.model.4.block.1.conv.conv""": """encoder.layers.4.block.1.conv""", """encoder.model.4.block.3.conv.conv""": """encoder.layers.4.block.3.conv""", """encoder.model.4.shortcut.conv.conv""": """encoder.layers.4.shortcut.conv""", """encoder.model.6.conv.conv""": """encoder.layers.6.conv""", """encoder.model.7.block.1.conv.conv""": """encoder.layers.7.block.1.conv""", """encoder.model.7.block.3.conv.conv""": """encoder.layers.7.block.3.conv""", """encoder.model.7.shortcut.conv.conv""": """encoder.layers.7.shortcut.conv""", """encoder.model.9.conv.conv""": """encoder.layers.9.conv""", """encoder.model.10.block.1.conv.conv""": """encoder.layers.10.block.1.conv""", """encoder.model.10.block.3.conv.conv""": """encoder.layers.10.block.3.conv""", """encoder.model.10.shortcut.conv.conv""": """encoder.layers.10.shortcut.conv""", """encoder.model.12.conv.conv""": """encoder.layers.12.conv""", """encoder.model.13.lstm""": """encoder.layers.13.lstm""", """encoder.model.15.conv.conv""": """encoder.layers.15.conv""", } __snake_case ={ """encoder.model.0.conv.norm""": """encoder.layers.0.norm""", """encoder.model.1.block.1.conv.norm""": """encoder.layers.1.block.1.norm""", """encoder.model.1.block.3.conv.norm""": """encoder.layers.1.block.3.norm""", """encoder.model.1.shortcut.conv.norm""": """encoder.layers.1.shortcut.norm""", """encoder.model.3.conv.norm""": """encoder.layers.3.norm""", """encoder.model.4.block.1.conv.norm""": """encoder.layers.4.block.1.norm""", """encoder.model.4.block.3.conv.norm""": """encoder.layers.4.block.3.norm""", """encoder.model.4.shortcut.conv.norm""": """encoder.layers.4.shortcut.norm""", """encoder.model.6.conv.norm""": """encoder.layers.6.norm""", """encoder.model.7.block.1.conv.norm""": """encoder.layers.7.block.1.norm""", """encoder.model.7.block.3.conv.norm""": """encoder.layers.7.block.3.norm""", """encoder.model.7.shortcut.conv.norm""": """encoder.layers.7.shortcut.norm""", """encoder.model.9.conv.norm""": """encoder.layers.9.norm""", """encoder.model.10.block.1.conv.norm""": """encoder.layers.10.block.1.norm""", """encoder.model.10.block.3.conv.norm""": """encoder.layers.10.block.3.norm""", """encoder.model.10.shortcut.conv.norm""": """encoder.layers.10.shortcut.norm""", """encoder.model.12.conv.norm""": """encoder.layers.12.norm""", """encoder.model.15.conv.norm""": """encoder.layers.15.norm""", } __snake_case ={ """decoder.model.0.conv.conv""": """decoder.layers.0.conv""", """decoder.model.1.lstm""": """decoder.layers.1.lstm""", """decoder.model.3.convtr.convtr""": """decoder.layers.3.conv""", """decoder.model.4.block.1.conv.conv""": """decoder.layers.4.block.1.conv""", """decoder.model.4.block.3.conv.conv""": """decoder.layers.4.block.3.conv""", """decoder.model.4.shortcut.conv.conv""": """decoder.layers.4.shortcut.conv""", """decoder.model.6.convtr.convtr""": """decoder.layers.6.conv""", """decoder.model.7.block.1.conv.conv""": """decoder.layers.7.block.1.conv""", """decoder.model.7.block.3.conv.conv""": """decoder.layers.7.block.3.conv""", """decoder.model.7.shortcut.conv.conv""": """decoder.layers.7.shortcut.conv""", """decoder.model.9.convtr.convtr""": """decoder.layers.9.conv""", """decoder.model.10.block.1.conv.conv""": """decoder.layers.10.block.1.conv""", """decoder.model.10.block.3.conv.conv""": """decoder.layers.10.block.3.conv""", """decoder.model.10.shortcut.conv.conv""": """decoder.layers.10.shortcut.conv""", """decoder.model.12.convtr.convtr""": """decoder.layers.12.conv""", """decoder.model.13.block.1.conv.conv""": """decoder.layers.13.block.1.conv""", """decoder.model.13.block.3.conv.conv""": """decoder.layers.13.block.3.conv""", """decoder.model.13.shortcut.conv.conv""": """decoder.layers.13.shortcut.conv""", """decoder.model.15.conv.conv""": """decoder.layers.15.conv""", } __snake_case ={ """decoder.model.0.conv.norm""": """decoder.layers.0.norm""", """decoder.model.3.convtr.norm""": """decoder.layers.3.norm""", """decoder.model.4.block.1.conv.norm""": """decoder.layers.4.block.1.norm""", """decoder.model.4.block.3.conv.norm""": """decoder.layers.4.block.3.norm""", """decoder.model.4.shortcut.conv.norm""": """decoder.layers.4.shortcut.norm""", """decoder.model.6.convtr.norm""": """decoder.layers.6.norm""", """decoder.model.7.block.1.conv.norm""": """decoder.layers.7.block.1.norm""", """decoder.model.7.block.3.conv.norm""": """decoder.layers.7.block.3.norm""", """decoder.model.7.shortcut.conv.norm""": """decoder.layers.7.shortcut.norm""", """decoder.model.9.convtr.norm""": """decoder.layers.9.norm""", """decoder.model.10.block.1.conv.norm""": """decoder.layers.10.block.1.norm""", """decoder.model.10.block.3.conv.norm""": """decoder.layers.10.block.3.norm""", """decoder.model.10.shortcut.conv.norm""": """decoder.layers.10.shortcut.norm""", """decoder.model.12.convtr.norm""": """decoder.layers.12.norm""", """decoder.model.13.block.1.conv.norm""": """decoder.layers.13.block.1.norm""", """decoder.model.13.block.3.conv.norm""": """decoder.layers.13.block.3.norm""", """decoder.model.13.shortcut.conv.norm""": """decoder.layers.13.shortcut.norm""", """decoder.model.15.conv.norm""": """decoder.layers.15.norm""", } __snake_case ={ **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } __snake_case ={ **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } __snake_case =[] __snake_case =[] def a_ ( lowerCamelCase : Optional[Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : int , lowerCamelCase : Any , lowerCamelCase : List[str] ): for attribute in key.split('.' ): lowerCAmelCase = getattr(lowerCamelCase , lowerCamelCase ) if weight_type is not None: lowerCAmelCase = getattr(lowerCamelCase , lowerCamelCase ).shape else: lowerCAmelCase = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' f''' {value.shape} for {full_name}''' ) if weight_type == "weight": lowerCAmelCase = value elif weight_type == "weight_g": lowerCAmelCase = value elif weight_type == "weight_v": lowerCAmelCase = value elif weight_type == "bias": lowerCAmelCase = value elif weight_type == "running_mean": lowerCAmelCase = value elif weight_type == "running_var": lowerCAmelCase = value elif weight_type == "num_batches_tracked": lowerCAmelCase = value elif weight_type == "weight_ih_l0": lowerCAmelCase = value elif weight_type == "weight_hh_l0": lowerCAmelCase = value elif weight_type == "bias_ih_l0": lowerCAmelCase = value elif weight_type == "bias_hh_l0": lowerCAmelCase = value elif weight_type == "weight_ih_l1": lowerCAmelCase = value elif weight_type == "weight_hh_l1": lowerCAmelCase = value elif weight_type == "bias_ih_l1": lowerCAmelCase = value elif weight_type == "bias_hh_l1": lowerCAmelCase = value else: lowerCAmelCase = value logger.info(f'''{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.''' ) def a_ ( lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ): for key in ignore_keys: if key.endswith('.*' ): if name.startswith(key[:-1] ): return True elif ".*." in key: lowerCAmelCase , lowerCAmelCase = key.split('.*.' ) if prefix in name and suffix in name: return True elif key in name: return True return False def a_ ( lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : str ): lowerCAmelCase = [] if model_name == "encodec_24khz" or "encodec_32khz": lowerCAmelCase = MAPPING_24K elif model_name == "encodec_48khz": lowerCAmelCase = MAPPING_48K else: raise ValueError(f'''Unsupported model: {model_name}''' ) for name, value in orig_dict.items(): if should_ignore(lowerCamelCase , lowerCamelCase ): logger.info(f'''{name} was ignored''' ) continue lowerCAmelCase = False for key, mapped_key in MAPPING.items(): if "*" in key: lowerCAmelCase , lowerCAmelCase = key.split('.*.' ) if prefix in name and suffix in name: lowerCAmelCase = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('embed' ) and name.endswith('embed_avg' ): continue lowerCAmelCase = True if "*" in mapped_key: lowerCAmelCase = name.split(lowerCamelCase )[0].split('.' )[-2] lowerCAmelCase = mapped_key.replace('*' , lowerCamelCase ) if "weight_g" in name: lowerCAmelCase = 'weight_g' elif "weight_v" in name: lowerCAmelCase = 'weight_v' elif "weight_ih_l0" in name: lowerCAmelCase = 'weight_ih_l0' elif "weight_hh_l0" in name: lowerCAmelCase = 'weight_hh_l0' elif "bias_ih_l0" in name: lowerCAmelCase = 'bias_ih_l0' elif "bias_hh_l0" in name: lowerCAmelCase = 'bias_hh_l0' elif "weight_ih_l1" in name: lowerCAmelCase = 'weight_ih_l1' elif "weight_hh_l1" in name: lowerCAmelCase = 'weight_hh_l1' elif "bias_ih_l1" in name: lowerCAmelCase = 'bias_ih_l1' elif "bias_hh_l1" in name: lowerCAmelCase = 'bias_hh_l1' elif "bias" in name: lowerCAmelCase = 'bias' elif "weight" in name: lowerCAmelCase = 'weight' elif "running_mean" in name: lowerCAmelCase = 'running_mean' elif "running_var" in name: lowerCAmelCase = 'running_var' elif "num_batches_tracked" in name: lowerCAmelCase = 'num_batches_tracked' else: lowerCAmelCase = None set_recursively(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) continue if not is_used: unused_weights.append(lowerCamelCase ) logger.warning(f'''Unused weights: {unused_weights}''' ) @torch.no_grad() def a_ ( lowerCamelCase : Optional[int] , lowerCamelCase : Union[str, Any] , lowerCamelCase : str , lowerCamelCase : Dict=None , lowerCamelCase : Union[str, Any]=None , ): if config_path is not None: lowerCAmelCase = EncodecConfig.from_pretrained(lowerCamelCase ) else: lowerCAmelCase = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": lowerCAmelCase = [8, 5, 4, 4] lowerCAmelCase = [2.2] lowerCAmelCase = 64 lowerCAmelCase = 32000 lowerCAmelCase = 2048 lowerCAmelCase = False lowerCAmelCase = False lowerCAmelCase = False elif model_name == "encodec_48khz": lowerCAmelCase = [8, 5, 4, 2] lowerCAmelCase = [3.0, 6.0, 12.0, 24.0] lowerCAmelCase = 48000 lowerCAmelCase = 2 lowerCAmelCase = False lowerCAmelCase = 'time_group_norm' lowerCAmelCase = True lowerCAmelCase = 1.0 lowerCAmelCase = 0.01 else: raise ValueError(f'''Unknown model name: {model_name}''' ) lowerCAmelCase = EncodecModel(lowerCamelCase ) lowerCAmelCase = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(lowerCamelCase ) lowerCAmelCase = torch.load(lowerCamelCase ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights lowerCAmelCase = original_checkpoint['best_state'] recursively_load_weights(lowerCamelCase , lowerCamelCase , lowerCamelCase ) model.save_pretrained(lowerCamelCase ) if repo_id: print('Pushing to the hub...' ) feature_extractor.push_to_hub(lowerCamelCase ) model.push_to_hub(lowerCamelCase ) if __name__ == "__main__": __snake_case =argparse.ArgumentParser() parser.add_argument( """--model""", default="""encodec_24khz""", type=str, help="""The model to convert. Should be one of 'encodec_24khz', 'encodec_32khz', 'encodec_48khz'.""", ) parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) __snake_case =parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
4
1
"""simple docstring""" from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class __lowerCamelCase ( a__ ): '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> Optional[Any]: super().__init__( __UpperCAmelCase , split=__UpperCAmelCase , features=__UpperCAmelCase , cache_dir=__UpperCAmelCase , keep_in_memory=__UpperCAmelCase , streaming=__UpperCAmelCase , num_proc=__UpperCAmelCase , **__UpperCAmelCase , ) _a = path_or_paths if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else {self.split: path_or_paths} _a = Text( cache_dir=__UpperCAmelCase , data_files=__UpperCAmelCase , features=__UpperCAmelCase , **__UpperCAmelCase , ) def _UpperCAmelCase ( self ) -> Any: # Build iterable dataset if self.streaming: _a = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _a = None _a = None _a = None _a = None self.builder.download_and_prepare( download_config=__UpperCAmelCase , download_mode=__UpperCAmelCase , verification_mode=__UpperCAmelCase , base_path=__UpperCAmelCase , num_proc=self.num_proc , ) _a = self.builder.as_dataset( split=self.split , verification_mode=__UpperCAmelCase , in_memory=self.keep_in_memory ) return dataset
365
"""simple docstring""" import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class __lowerCamelCase : '''simple docstring''' @staticmethod def _UpperCAmelCase ( *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: pass def A_ ( _lowerCAmelCase : Image ): """simple docstring""" _a = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' A_ : List[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> str: _a = DepthEstimationPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Dict: _a = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , __UpperCAmelCase ) import datasets _a = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) _a = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , __UpperCAmelCase , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def _UpperCAmelCase ( self ) -> Tuple: pass @slow @require_torch def _UpperCAmelCase ( self ) -> List[str]: _a = '''Intel/dpt-large''' _a = pipeline('''depth-estimation''' , model=__UpperCAmelCase ) _a = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) _a = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 29.304 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.662 ) @require_torch def _UpperCAmelCase ( self ) -> List[Any]: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
153
0
import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py lowerCAmelCase : Optional[Any] = """src/diffusers""" # Matches is_xxx_available() lowerCAmelCase : str = re.compile(R"""is\_([a-z_]*)_available\(\)""") # Matches from xxx import bla lowerCAmelCase : List[str] = re.compile(R"""\s+from\s+\S*\s+import\s+([^\(\s].*)\n""") lowerCAmelCase : Union[str, Any] = """ {0} = None """ lowerCAmelCase : List[str] = """ class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) @classmethod def from_config(cls, *args, **kwargs): requires_backends(cls, {1}) @classmethod def from_pretrained(cls, *args, **kwargs): requires_backends(cls, {1}) """ lowerCAmelCase : List[Any] = """ def {0}(*args, **kwargs): requires_backends({0}, {1}) """ def A_ ( _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = _re_backend.findall(_UpperCAmelCase ) if len(_UpperCAmelCase ) == 0: return None return "_and_".join(_UpperCAmelCase ) def A_ ( ): with open(os.path.join(_UpperCAmelCase , "__init__.py" ) , "r" , encoding="utf-8" , newline="\n" ) as f: SCREAMING_SNAKE_CASE_: Optional[Any] = f.readlines() # Get to the point we do the actual imports for type checking SCREAMING_SNAKE_CASE_: Union[str, Any] = 0 SCREAMING_SNAKE_CASE_: Optional[int] = {} # Go through the end of the file while line_index < len(_UpperCAmelCase ): # If the line contains is_backend_available, we grab all objects associated with the `else` block SCREAMING_SNAKE_CASE_: List[str] = find_backend(lines[line_index] ) if backend is not None: while not lines[line_index].startswith("else:" ): line_index += 1 line_index += 1 SCREAMING_SNAKE_CASE_: str = [] # Until we unindent, add backend objects to the list while line_index < len(_UpperCAmelCase ) and len(lines[line_index] ) > 1: SCREAMING_SNAKE_CASE_: List[Any] = lines[line_index] SCREAMING_SNAKE_CASE_: Optional[int] = _re_single_line_import.search(_UpperCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 8 ): objects.append(line[8:-2] ) line_index += 1 if len(_UpperCAmelCase ) > 0: SCREAMING_SNAKE_CASE_: Any = objects else: line_index += 1 return backend_specific_objects def A_ ( _UpperCAmelCase , _UpperCAmelCase ): if name.isupper(): return DUMMY_CONSTANT.format(_UpperCAmelCase ) elif name.islower(): return DUMMY_FUNCTION.format(_UpperCAmelCase , _UpperCAmelCase ) else: return DUMMY_CLASS.format(_UpperCAmelCase , _UpperCAmelCase ) def A_ ( _UpperCAmelCase=None ): if backend_specific_objects is None: SCREAMING_SNAKE_CASE_: Optional[Any] = read_init() # For special correspondence backend to module name as used in the function requires_modulename SCREAMING_SNAKE_CASE_: Optional[Any] = {} for backend, objects in backend_specific_objects.items(): SCREAMING_SNAKE_CASE_: Tuple = "[" + ", ".join(f"\"{b}\"" for b in backend.split("_and_" ) ) + "]" SCREAMING_SNAKE_CASE_: Tuple = "# This file is autogenerated by the command `make fix-copies`, do not edit.\n" dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(_UpperCAmelCase , _UpperCAmelCase ) for o in objects] ) SCREAMING_SNAKE_CASE_: Any = dummy_file return dummy_files def A_ ( _UpperCAmelCase=False ): SCREAMING_SNAKE_CASE_: Union[str, Any] = create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py SCREAMING_SNAKE_CASE_: int = {"torch": "pt"} # Locate actual dummy modules and read their content. SCREAMING_SNAKE_CASE_: Dict = os.path.join(_UpperCAmelCase , "utils" ) SCREAMING_SNAKE_CASE_: List[Any] = { backend: os.path.join(_UpperCAmelCase , f"dummy_{short_names.get(_UpperCAmelCase , _UpperCAmelCase )}_objects.py" ) for backend in dummy_files.keys() } SCREAMING_SNAKE_CASE_: Any = {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(_UpperCAmelCase ): with open(_UpperCAmelCase , "r" , encoding="utf-8" , newline="\n" ) as f: SCREAMING_SNAKE_CASE_: Tuple = f.read() else: SCREAMING_SNAKE_CASE_: Optional[Any] = "" for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( f"Updating diffusers.utils.dummy_{short_names.get(_UpperCAmelCase , _UpperCAmelCase )}_objects.py as the main " "__init__ has new objects." ) with open(dummy_file_paths[backend] , "w" , encoding="utf-8" , newline="\n" ) as f: f.write(dummy_files[backend] ) else: raise ValueError( "The main __init__ has objects that are not present in " f"diffusers.utils.dummy_{short_names.get(_UpperCAmelCase , _UpperCAmelCase )}_objects.py. Run `make fix-copies` " "to fix this." ) if __name__ == "__main__": lowerCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") lowerCAmelCase : str = parser.parse_args() check_dummies(args.fix_and_overwrite)
13
"""simple docstring""" import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm lowercase_ = re.compile("[^A-Za-z_0-9]") # parameters used in DuplicationIndex lowercase_ = 1_0 lowercase_ = 2_5_6 def lowercase ( lowerCAmelCase__ : List[str] ) -> Optional[MinHash]: if len(lowerCAmelCase__ ) < MIN_NUM_TOKENS: return None __a = MinHash(num_perm=lowerCAmelCase__ ) for token in set(lowerCAmelCase__ ): min_hash.update(token.encode() ) return min_hash def lowercase ( lowerCAmelCase__ : str ) -> Set[str]: return {t for t in NON_ALPHA.split(lowerCAmelCase__ ) if len(t.strip() ) > 0} class __lowerCAmelCase : '''simple docstring''' def __init__( self , *, _a = 0.85 , ): __a = duplication_jaccard_threshold __a = NUM_PERM __a = MinHashLSH(threshold=self._duplication_jaccard_threshold , num_perm=self._num_perm ) __a = defaultdict(_a ) def __UpperCAmelCase ( self , _a , _a ): __a = self._index.query(_a ) if code_key in self._index.keys: print(f'''Duplicate key {code_key}''' ) return self._index.insert(_a , _a ) if len(_a ) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(_a ) break else: self._duplicate_clusters[close_duplicates[0]].add(_a ) def __UpperCAmelCase ( self ): __a = [] for base, duplicates in self._duplicate_clusters.items(): __a = [base] + list(_a ) # reformat the cluster to be a list of dict __a = [{'''base_index''': el[0], '''repo_name''': el[1], '''path''': el[2]} for el in cluster] duplicate_clusters.append(_a ) return duplicate_clusters def __UpperCAmelCase ( self , _a ): __a = self.get_duplicate_clusters() with open(_a , '''w''' ) as f: json.dump(_a , _a ) def lowercase ( lowerCAmelCase__ : List[str] ) -> int: __a , __a = element __a = get_min_hash([t for t in NON_ALPHA.split(data['''content'''] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def lowercase ( lowerCAmelCase__ : Type[Dataset] ) -> str: with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(lowerCAmelCase__ , max_queue_size=10000 ) , chunksize=100 , ): if data is not None: yield data def lowercase ( lowerCAmelCase__ : Type[Dataset] , lowerCAmelCase__ : float ) -> Dict: __a = DuplicationIndex(duplication_jaccard_threshold=lowerCAmelCase__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(lowerCAmelCase__ ) ) , max_queue_size=100 ) ): di.add(lowerCAmelCase__ , lowerCAmelCase__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def lowercase ( lowerCAmelCase__ : str , lowerCAmelCase__ : str ) -> float: __a = get_tokens(lowerCAmelCase__ ) __a = get_tokens(lowerCAmelCase__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) lowercase_ = None def lowercase ( lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Union[str, Any] ) -> Any: __a = [] for elementa in cluster: __a = _shared_dataset[elementa['''base_index''']]['''content'''] for elementa in extremes: __a = _shared_dataset[elementa['''base_index''']]['''content'''] if jaccard_similarity(lowerCAmelCase__ , lowerCAmelCase__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: __a = 1 extremes.append(lowerCAmelCase__ ) return extremes def lowercase ( lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[int] ) -> Optional[int]: global _shared_dataset __a = dataset __a = [] __a = partial(_find_cluster_extremes_shared , jaccard_threshold=lowerCAmelCase__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( lowerCAmelCase__ , lowerCAmelCase__ , ) , total=len(lowerCAmelCase__ ) , ): extremes_list.append(lowerCAmelCase__ ) return extremes_list def lowercase ( lowerCAmelCase__ : Type[Dataset] , lowerCAmelCase__ : float = 0.85 ) -> Tuple[Type[Dataset], List[List[Dict]]]: __a = make_duplicate_clusters(lowerCAmelCase__ , lowerCAmelCase__ ) __a = {x['''base_index'''] for cluster in duplicate_clusters for x in cluster} __a = {} __a = find_extremes(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for extremes in extremes_clusters: for element in extremes: __a = element __a = duplicate_indices - set(extreme_dict.keys() ) __a = dataset.filter(lambda lowerCAmelCase__ , lowerCAmelCase__ : idx not in remove_indices , with_indices=lowerCAmelCase__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: __a = element['''base_index'''] in extreme_dict if element["is_extreme"]: __a = extreme_dict[element['''base_index''']]['''copies'''] print(f'''Original dataset size: {len(lowerCAmelCase__ )}''' ) print(f'''Number of duplicate clusters: {len(lowerCAmelCase__ )}''' ) print(f'''Files in duplicate cluster: {len(lowerCAmelCase__ )}''' ) print(f'''Unique files in duplicate cluster: {len(lowerCAmelCase__ )}''' ) print(f'''Filtered dataset size: {len(lowerCAmelCase__ )}''' ) return ds_filter, duplicate_clusters
45
0
"""simple docstring""" def __lowerCamelCase ( __UpperCamelCase ) -> int: """simple docstring""" lowerCAmelCase_ : List[Any] = hex_num.strip() if not hex_num: raise ValueError("No value was passed to the function" ) lowerCAmelCase_ : Tuple = hex_num[0] == "-" if is_negative: lowerCAmelCase_ : Optional[int] = hex_num[1:] try: lowerCAmelCase_ : Optional[int] = int(__UpperCamelCase , 16 ) except ValueError: raise ValueError("Invalid value was passed to the function" ) lowerCAmelCase_ : Optional[int] = "" while int_num > 0: lowerCAmelCase_ : List[str] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(("-" + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
357
"""simple docstring""" class __lowerCamelCase ( A__ ): '''simple docstring''' pass class __lowerCamelCase ( A__ ): '''simple docstring''' pass class __lowerCamelCase : '''simple docstring''' def __init__( self : List[str] ): lowerCAmelCase_ : Optional[int] = [ [], [], [], ] def lowerCamelCase ( self : List[str] , a_ : int , a_ : int ): try: if len(self.queues[priority] ) >= 1_00: raise OverflowError("Maximum queue size is 100" ) self.queues[priority].append(a_ ) except IndexError: raise ValueError("Valid priorities are 0, 1, and 2" ) def lowerCamelCase ( self : Dict ): for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("All queues are empty" ) def __str__( self : Any ): return "\n".join(f'''Priority {i}: {q}''' for i, q in enumerate(self.queues ) ) class __lowerCamelCase : '''simple docstring''' def __init__( self : List[Any] ): lowerCAmelCase_ : int = [] def lowerCamelCase ( self : int , a_ : int ): if len(self.queue ) == 1_00: raise OverFlowError("Maximum queue size is 100" ) self.queue.append(a_ ) def lowerCamelCase ( self : List[str] ): if not self.queue: raise UnderFlowError("The queue is empty" ) else: lowerCAmelCase_ : Optional[int] = min(self.queue ) self.queue.remove(a_ ) return data def __str__( self : List[str] ): return str(self.queue ) def __lowerCamelCase ( ) -> List[str]: """simple docstring""" lowerCAmelCase_ : Tuple = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(__UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(__UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def __lowerCamelCase ( ) -> str: """simple docstring""" lowerCAmelCase_ : Union[str, Any] = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(__UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(__UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
161
0
"""simple docstring""" def __lowerCAmelCase ( lowercase : int = 100 ) -> Optional[Any]: """simple docstring""" snake_case : Tuple = (n * (n + 1) // 2) ** 2 snake_case : Tuple = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(F'''{solution() = }''')
203
import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase__ = get_tests_dir("""fixtures/test_sentencepiece.model""") UpperCamelCase__ = get_tests_dir("""fixtures/test_sentencepiece_bpe.model""") UpperCamelCase__ = """pt""" if is_torch_available() else """tf""" @require_sentencepiece @require_tokenizers class a__ ( snake_case__ , unittest.TestCase ): _a : int = CamembertTokenizer _a : Dict = CamembertTokenizerFast _a : Tuple = True _a : List[Any] = True def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __lowerCAmelCase = CamembertTokenizer(_A ) tokenizer.save_pretrained(self.tmpdirname ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = "<pad>" __lowerCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>NOTUSED" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(_A ) , 1_0_0_4 ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_5 ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = CamembertTokenizer(_A ) tokenizer.save_pretrained(self.tmpdirname ) __lowerCAmelCase = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) __lowerCAmelCase = "I was born in 92000, and this is falsé." __lowerCAmelCase = tokenizer.encode(_A ) __lowerCAmelCase = rust_tokenizer.encode(_A ) self.assertListEqual(_A , _A ) __lowerCAmelCase = tokenizer.encode(_A , add_special_tokens=_A ) __lowerCAmelCase = rust_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) __lowerCAmelCase = tokenizer.convert_ids_to_tokens(_A ) __lowerCAmelCase = rust_tokenizer.tokenize(_A ) self.assertListEqual(_A , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if not self.test_rust_tokenizer: return __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = self.get_rust_tokenizer() __lowerCAmelCase = "I was born in 92000, and this is falsé." __lowerCAmelCase = tokenizer.tokenize(_A ) __lowerCAmelCase = rust_tokenizer.tokenize(_A ) self.assertListEqual(_A , _A ) __lowerCAmelCase = tokenizer.encode(_A , add_special_tokens=_A ) __lowerCAmelCase = rust_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) __lowerCAmelCase = self.get_rust_tokenizer() __lowerCAmelCase = tokenizer.encode(_A ) __lowerCAmelCase = rust_tokenizer.encode(_A ) self.assertListEqual(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = {"input_ids": [[5, 5_4, 7_1_9_6, 2_9_7, 3_0, 2_3, 7_7_6, 1_8, 1_1, 3_2_1_5, 3_7_0_5, 8_2_5_2, 2_2, 3_1_6_4, 1_1_8_1, 2_1_1_6, 2_9, 1_6, 8_1_3, 2_5, 7_9_1, 3_3_1_4, 2_0, 3_4_4_6, 3_8, 2_7_5_7_5, 1_2_0, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 4_6_8, 1_7, 1_1, 9_0_8_8, 2_0, 1_5_1_7, 8, 2_2_8_0_4, 1_8_8_1_8, 1_0, 3_8, 6_2_9, 6_0_7, 6_0_7, 1_4_2, 1_9, 7_1_9_6, 8_6_7, 5_6, 1_0_3_2_6, 2_4, 2_2_6_7, 2_0, 4_1_6, 5_0_7_2, 1_5_6_1_2, 2_3_3, 7_3_4, 7, 2_3_9_9, 2_7, 1_6, 3_0_1_5, 1_6_4_9, 7, 2_4, 2_0, 4_3_3_8, 2_3_9_9, 2_7, 1_3, 3_4_0_0, 1_4, 1_3, 6_1_8_9, 8, 9_3_0, 9, 6]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. __lowerCAmelCase = [ "Le transformeur est un modèle d'apprentissage profond introduit en 2017, " "utilisé principalement dans le domaine du traitement automatique des langues (TAL).", "À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus " "pour gérer des données séquentielles, telles que le langage naturel, pour des tâches " "telles que la traduction et la synthèse de texte.", ] self.tokenizer_integration_test_util( expected_encoding=_A , model_name="camembert-base" , revision="3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf" , sequences=_A , )
92
0
from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class UpperCamelCase_ : '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : Tuple , UpperCAmelCase__ : Optional[Any]) ->List[Any]: '''simple docstring''' raise NotImplementedError() def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Optional[int]: '''simple docstring''' raise NotImplementedError() class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self : Tuple , UpperCAmelCase__ : "AutoTokenizer" , UpperCAmelCase__ : bool = False , **UpperCAmelCase__ : str) ->int: '''simple docstring''' A__ = tokenizer A__ = skip_prompt A__ = decode_kwargs # variables used in the streaming process A__ = [] A__ = 0 A__ = True def SCREAMING_SNAKE_CASE ( self : int , UpperCAmelCase__ : Any) ->Optional[Any]: '''simple docstring''' if len(value.shape) > 1 and value.shape[0] > 1: raise ValueError('''TextStreamer only supports batch size 1''') elif len(value.shape) > 1: A__ = value[0] if self.skip_prompt and self.next_tokens_are_prompt: A__ = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist()) A__ = self.tokenizer.decode(self.token_cache , **self.decode_kwargs) # After the symbol for a new line, we flush the cache. if text.endswith('''\n'''): A__ = text[self.print_len :] A__ = [] A__ = 0 # If the last token is a CJK character, we print the characters. elif len(UpperCAmelCase__) > 0 and self._is_chinese_char(ord(text[-1])): A__ = text[self.print_len :] self.print_len += len(UpperCAmelCase__) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: A__ = text[self.print_len : text.rfind(''' ''') + 1] self.print_len += len(UpperCAmelCase__) self.on_finalized_text(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Tuple: '''simple docstring''' if len(self.token_cache) > 0: A__ = self.tokenizer.decode(self.token_cache , **self.decode_kwargs) A__ = text[self.print_len :] A__ = [] A__ = 0 else: A__ = '''''' A__ = True self.on_finalized_text(UpperCAmelCase__ , stream_end=UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : bool = False) ->Dict: '''simple docstring''' print(UpperCAmelCase__ , flush=UpperCAmelCase__ , end='''''' if not stream_end else None) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : List[Any]) ->Union[str, Any]: '''simple docstring''' if ( (cp >= 0x4e_00 and cp <= 0x9f_ff) or (cp >= 0x34_00 and cp <= 0x4d_bf) # or (cp >= 0x2_00_00 and cp <= 0x2_a6_df) # or (cp >= 0x2_a7_00 and cp <= 0x2_b7_3f) # or (cp >= 0x2_b7_40 and cp <= 0x2_b8_1f) # or (cp >= 0x2_b8_20 and cp <= 0x2_ce_af) # or (cp >= 0xf9_00 and cp <= 0xfa_ff) or (cp >= 0x2_f8_00 and cp <= 0x2_fa_1f) # ): # return True return False class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self : Optional[Any] , UpperCAmelCase__ : "AutoTokenizer" , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : Optional[float] = None , **UpperCAmelCase__ : int) ->List[Any]: '''simple docstring''' super().__init__(UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__) A__ = Queue() A__ = None A__ = timeout def SCREAMING_SNAKE_CASE ( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : bool = False) ->int: '''simple docstring''' self.text_queue.put(UpperCAmelCase__ , timeout=self.timeout) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout) def __iter__( self : List[Any]) ->Tuple: '''simple docstring''' return self def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->Dict: '''simple docstring''' A__ = self.text_queue.get(timeout=self.timeout) if value == self.stop_signal: raise StopIteration() else: return value
352
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def SCREAMING_SNAKE_CASE ( lowercase_ = "" ) -> dict[str, float]: """simple docstring""" A__ = url or '''https://www.imdb.com/chart/top/?ref_=nv_mv_250''' A__ = BeautifulSoup(requests.get(lowercase_ ).text , '''html.parser''' ) A__ = soup.find_all('''td''' , attrs='''titleColumn''' ) A__ = soup.find_all('''td''' , class_='''ratingColumn imdbRating''' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(lowercase_ , lowercase_ ) } def SCREAMING_SNAKE_CASE ( lowercase_ = "IMDb_Top_250_Movies.csv" ) -> None: """simple docstring""" A__ = get_imdb_top_aaa_movies() with open(lowercase_ , '''w''' , newline='''''' ) as out_file: A__ = csv.writer(lowercase_ ) writer.writerow(['''Movie title''', '''IMDb rating'''] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
231
0
import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import OpenLlamaForCausalLM, OpenLlamaForSequenceClassification, OpenLlamaModel class __snake_case : def __init__( self : str , _snake_case : Tuple , _snake_case : List[str]=13 , _snake_case : Tuple=7 , _snake_case : Optional[Any]=True , _snake_case : List[str]=True , _snake_case : Dict=False , _snake_case : List[str]=True , _snake_case : Optional[int]=99 , _snake_case : Optional[int]=32 , _snake_case : Dict=5 , _snake_case : List[Any]=4 , _snake_case : Optional[int]=37 , _snake_case : Optional[Any]="gelu" , _snake_case : Tuple=0.1 , _snake_case : Tuple=0.1 , _snake_case : int=512 , _snake_case : int=16 , _snake_case : List[Any]=2 , _snake_case : Any=0.0_2 , _snake_case : str=3 , _snake_case : Union[str, Any]=4 , _snake_case : str=None , ): """simple docstring""" UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_input_mask UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = num_labels UpperCAmelCase_ = num_choices UpperCAmelCase_ = scope def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length]) UpperCAmelCase_ = None if self.use_token_type_ids: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices) UpperCAmelCase_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase ( self : str): """simple docstring""" return OpenLlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_snake_case , initializer_range=self.initializer_range , use_stable_embedding=_snake_case , ) def lowerCamelCase ( self : Optional[Any] , _snake_case : int , _snake_case : str , _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : Optional[Any] , _snake_case : Any , _snake_case : Optional[Any]): """simple docstring""" UpperCAmelCase_ = OpenLlamaModel(config=_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case , attention_mask=_snake_case) UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def lowerCamelCase ( self : Dict , _snake_case : Optional[int] , _snake_case : Dict , _snake_case : List[Any] , _snake_case : List[Any] , _snake_case : Any , _snake_case : List[Any] , _snake_case : str , _snake_case : Optional[int] , _snake_case : str , ): """simple docstring""" UpperCAmelCase_ = True UpperCAmelCase_ = OpenLlamaModel(_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model( _snake_case , attention_mask=_snake_case , encoder_hidden_states=_snake_case , encoder_attention_mask=_snake_case , ) UpperCAmelCase_ = model( _snake_case , attention_mask=_snake_case , encoder_hidden_states=_snake_case , ) UpperCAmelCase_ = model(_snake_case , attention_mask=_snake_case) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def lowerCamelCase ( self : Union[str, Any] , _snake_case : List[Any] , _snake_case : Tuple , _snake_case : int , _snake_case : List[Any] , _snake_case : Any , _snake_case : str , _snake_case : int , _snake_case : Optional[int] , _snake_case : Union[str, Any] , ): """simple docstring""" UpperCAmelCase_ = OpenLlamaForCausalLM(config=_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case , attention_mask=_snake_case , labels=_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def lowerCamelCase ( self : List[str] , _snake_case : Union[str, Any] , _snake_case : Optional[int] , _snake_case : Any , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : str , _snake_case : Tuple , _snake_case : Tuple , _snake_case : Dict , ): """simple docstring""" UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = OpenLlamaForCausalLM(config=_snake_case) model.to(_snake_case) model.eval() # first forward pass UpperCAmelCase_ = model( _snake_case , attention_mask=_snake_case , encoder_hidden_states=_snake_case , encoder_attention_mask=_snake_case , use_cache=_snake_case , ) UpperCAmelCase_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size) UpperCAmelCase_ = ids_tensor((self.batch_size, 3) , vocab_size=2) # append to next input_ids and UpperCAmelCase_ = torch.cat([input_ids, next_tokens] , dim=-1) UpperCAmelCase_ = torch.cat([input_mask, next_mask] , dim=-1) UpperCAmelCase_ = model( _snake_case , attention_mask=_snake_case , encoder_hidden_states=_snake_case , encoder_attention_mask=_snake_case , output_hidden_states=_snake_case , )['''hidden_states'''][0] UpperCAmelCase_ = model( _snake_case , attention_mask=_snake_case , encoder_hidden_states=_snake_case , encoder_attention_mask=_snake_case , past_key_values=_snake_case , output_hidden_states=_snake_case , )['''hidden_states'''][0] # select random slice UpperCAmelCase_ = ids_tensor((1,) , output_from_past.shape[-1]).item() UpperCAmelCase_ = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_snake_case , _snake_case , atol=1e-3)) def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __snake_case ( a , a , a , unittest.TestCase ): UpperCAmelCase__ : Dict = ( (OpenLlamaModel, OpenLlamaForCausalLM, OpenLlamaForSequenceClassification) if is_torch_available() else () ) UpperCAmelCase__ : int = (OpenLlamaForCausalLM,) if is_torch_available() else () UpperCAmelCase__ : Union[str, Any] = ( { '''feature-extraction''': OpenLlamaModel, '''text-classification''': OpenLlamaForSequenceClassification, '''text-generation''': OpenLlamaForCausalLM, '''zero-shot''': OpenLlamaForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : Any = False UpperCAmelCase__ : Dict = False def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = OpenLlamaModelTester(self) UpperCAmelCase_ = ConfigTester(self , config_class=_snake_case , hidden_size=37) def lowerCamelCase ( self : Tuple): """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase_ = type self.model_tester.create_and_check_model(*_snake_case) def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = 3 UpperCAmelCase_ = input_dict['''input_ids'''] UpperCAmelCase_ = input_ids.ne(1).to(_snake_case) UpperCAmelCase_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) UpperCAmelCase_ = OpenLlamaForSequenceClassification(_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case , attention_mask=_snake_case , labels=_snake_case) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = 3 UpperCAmelCase_ = '''single_label_classification''' UpperCAmelCase_ = input_dict['''input_ids'''] UpperCAmelCase_ = input_ids.ne(1).to(_snake_case) UpperCAmelCase_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) UpperCAmelCase_ = OpenLlamaForSequenceClassification(_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case , attention_mask=_snake_case , labels=_snake_case) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = 3 UpperCAmelCase_ = '''multi_label_classification''' UpperCAmelCase_ = input_dict['''input_ids'''] UpperCAmelCase_ = input_ids.ne(1).to(_snake_case) UpperCAmelCase_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) UpperCAmelCase_ = OpenLlamaForSequenceClassification(_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case , attention_mask=_snake_case , labels=_snake_case) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @unittest.skip('''Open-Llama buffers include complex numbers, which breaks this test''') def lowerCamelCase ( self : List[Any]): """simple docstring""" pass @parameterized.expand([('''linear''',), ('''dynamic''',)]) def lowerCamelCase ( self : List[Any] , _snake_case : List[Any]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = ids_tensor([1, 10] , config.vocab_size) UpperCAmelCase_ = ids_tensor([1, int(config.max_position_embeddings * 1.5)] , config.vocab_size) set_seed(42) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase_ = OpenLlamaModel(_snake_case) original_model.to(_snake_case) original_model.eval() UpperCAmelCase_ = original_model(_snake_case).last_hidden_state UpperCAmelCase_ = original_model(_snake_case).last_hidden_state set_seed(42) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase_ = {'''type''': scaling_type, '''factor''': 1_0.0} UpperCAmelCase_ = OpenLlamaModel(_snake_case) scaled_model.to(_snake_case) scaled_model.eval() UpperCAmelCase_ = scaled_model(_snake_case).last_hidden_state UpperCAmelCase_ = scaled_model(_snake_case).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(_snake_case , _snake_case , atol=1e-5)) else: self.assertFalse(torch.allclose(_snake_case , _snake_case , atol=1e-5)) # The output should be different for long inputs self.assertFalse(torch.allclose(_snake_case , _snake_case , atol=1e-5))
51
'''simple docstring''' import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def lowercase ( __magic_name__ , __magic_name__=10 ): '''simple docstring''' UpperCAmelCase : Tuple = [] for _ in range(__magic_name__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def lowercase ( __magic_name__ , __magic_name__=10 ): '''simple docstring''' UpperCAmelCase : List[str] = [] for step in range(__magic_name__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase : Any = os.path.join(__magic_name__ , "schedule.bin" ) torch.save(scheduler.state_dict() , __magic_name__ ) UpperCAmelCase : Any = torch.load(__magic_name__ ) scheduler.load_state_dict(__magic_name__ ) return lrs @require_torch class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def A_ ( self , snake_case , snake_case , snake_case ): '''simple docstring''' self.assertEqual(len(snake_case ) , len(snake_case ) ) for a, b in zip(snake_case , snake_case ): self.assertAlmostEqual(snake_case , snake_case , delta=snake_case ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : Dict = torch.tensor([0.1, -0.2, -0.1] , requires_grad=snake_case ) UpperCAmelCase : Any = torch.tensor([0.4, 0.2, -0.5] ) UpperCAmelCase : Any = nn.MSELoss() # No warmup, constant schedule, no gradient clipping UpperCAmelCase : List[str] = AdamW(params=[w] , lr=2e-1 , weight_decay=0.0 ) for _ in range(1_0_0 ): UpperCAmelCase : List[Any] = criterion(snake_case , snake_case ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : Tuple = torch.tensor([0.1, -0.2, -0.1] , requires_grad=snake_case ) UpperCAmelCase : int = torch.tensor([0.4, 0.2, -0.5] ) UpperCAmelCase : str = nn.MSELoss() # No warmup, constant schedule, no gradient clipping UpperCAmelCase : str = Adafactor( params=[w] , lr=1e-2 , eps=(1e-30, 1e-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=snake_case , weight_decay=0.0 , relative_step=snake_case , scale_parameter=snake_case , warmup_init=snake_case , ) for _ in range(1_0_0_0 ): UpperCAmelCase : str = criterion(snake_case , snake_case ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) @require_torch class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = nn.Linear(50 , 50 ) if is_torch_available() else None SCREAMING_SNAKE_CASE__ : List[Any] = AdamW(m.parameters() , lr=1_0.0 ) if is_torch_available() else None SCREAMING_SNAKE_CASE__ : Optional[int] = 10 def A_ ( self , snake_case , snake_case , snake_case , snake_case=None ): '''simple docstring''' self.assertEqual(len(snake_case ) , len(snake_case ) ) for a, b in zip(snake_case , snake_case ): self.assertAlmostEqual(snake_case , snake_case , delta=snake_case , msg=snake_case ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : int = {"num_warmup_steps": 2, "num_training_steps": 1_0} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) UpperCAmelCase : int = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1e-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): UpperCAmelCase , UpperCAmelCase : Any = data UpperCAmelCase : Tuple = scheduler_func(self.optimizer , **snake_case ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) UpperCAmelCase : List[str] = unwrap_schedule(snake_case , self.num_steps ) self.assertListAlmostEqual( snake_case , snake_case , tol=1e-2 , msg=f"failed for {scheduler_func} in normal scheduler" , ) UpperCAmelCase : Optional[Any] = scheduler_func(self.optimizer , **snake_case ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(snake_case ) # wrap to test picklability of the schedule UpperCAmelCase : Tuple = unwrap_and_save_reload_schedule(snake_case , self.num_steps ) self.assertListEqual(snake_case , snake_case , msg=f"failed for {scheduler_func} in save and reload" ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case ): '''simple docstring''' UpperCAmelCase : List[str] = fn def __call__( self , *snake_case , **snake_case ): '''simple docstring''' return self.fn(*snake_case , **snake_case ) @classmethod def A_ ( self , snake_case ): '''simple docstring''' UpperCAmelCase : Optional[int] = list(map(self , scheduler.lr_lambdas ) )
311
0
"""simple docstring""" A : Union[str, Any] = 8.3_144_598 def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example A : Optional[Any] = 3_0_0 A : str = 2_8 A : Dict = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
259
"""simple docstring""" def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = [False] * len(_UpperCamelCase ) __lowerCAmelCase = [] queue.append(_UpperCamelCase ) __lowerCAmelCase = True while queue: __lowerCAmelCase = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_UpperCamelCase ) __lowerCAmelCase = True __lowerCAmelCase = u return visited[t] def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = [-1] * (len(_UpperCamelCase )) __lowerCAmelCase = 0 while bfs(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase = float("Inf" ) __lowerCAmelCase = sink while s != source: # Find the minimum value in select path __lowerCAmelCase = min(_UpperCamelCase , graph[parent[s]][s] ) __lowerCAmelCase = parent[s] max_flow += path_flow __lowerCAmelCase = sink while v != source: __lowerCAmelCase = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow __lowerCAmelCase = parent[v] return max_flow A : Optional[Any] = [ [0, 1_6, 1_3, 0, 0, 0], [0, 0, 1_0, 1_2, 0, 0], [0, 4, 0, 0, 1_4, 0], [0, 0, 9, 0, 0, 2_0], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] A , A : Optional[Any] = 0, 5 print(ford_fulkerson(graph, source, sink))
259
1
'''simple docstring''' lowerCamelCase : Any = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" lowerCamelCase : int = [{"type": "code", "content": INSTALL_CONTENT}] lowerCamelCase : str = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
47
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def _lowerCAmelCase ( _UpperCamelCase : Optional[int] ) -> List[str]: """simple docstring""" _SCREAMING_SNAKE_CASE =botoa.client('iam' ) _SCREAMING_SNAKE_CASE ={ 'Version': '2012-10-17', 'Statement': [ {'Effect': 'Allow', 'Principal': {'Service': 'sagemaker.amazonaws.com'}, 'Action': 'sts:AssumeRole'} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=_UpperCamelCase , AssumeRolePolicyDocument=json.dumps(_UpperCamelCase , indent=2 ) ) _SCREAMING_SNAKE_CASE ={ 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 'sagemaker:*', 'ecr:GetDownloadUrlForLayer', 'ecr:BatchGetImage', 'ecr:BatchCheckLayerAvailability', 'ecr:GetAuthorizationToken', 'cloudwatch:PutMetricData', 'cloudwatch:GetMetricData', 'cloudwatch:GetMetricStatistics', 'cloudwatch:ListMetrics', 'logs:CreateLogGroup', 'logs:CreateLogStream', 'logs:DescribeLogStreams', 'logs:PutLogEvents', 'logs:GetLogEvents', 's3:CreateBucket', 's3:ListBucket', 's3:GetBucketLocation', 's3:GetObject', 's3:PutObject', ], 'Resource': '*', } ], } # attach policy to role iam_client.put_role_policy( RoleName=_UpperCamelCase , PolicyName=f"{role_name}_policy_permission" , PolicyDocument=json.dumps(_UpperCamelCase , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(f"role {role_name} already exists. Using existing one" ) def _lowerCAmelCase ( _UpperCamelCase : List[str] ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =botoa.client('iam' ) return iam_client.get_role(RoleName=_UpperCamelCase )["Role"]["Arn"] def _lowerCAmelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =_ask_options( 'How do you want to authorize?' , ['AWS Profile', 'Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) '] , _UpperCamelCase , ) _SCREAMING_SNAKE_CASE =None if credentials_configuration == 0: _SCREAMING_SNAKE_CASE =_ask_field('Enter your AWS Profile name: [default] ' , default='default' ) _SCREAMING_SNAKE_CASE =aws_profile else: print( 'Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,' '`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`' ) _SCREAMING_SNAKE_CASE =_ask_field('AWS Access Key ID: ' ) _SCREAMING_SNAKE_CASE =aws_access_key_id _SCREAMING_SNAKE_CASE =_ask_field('AWS Secret Access Key: ' ) _SCREAMING_SNAKE_CASE =aws_secret_access_key _SCREAMING_SNAKE_CASE =_ask_field('Enter your AWS Region: [us-east-1]' , default='us-east-1' ) _SCREAMING_SNAKE_CASE =aws_region _SCREAMING_SNAKE_CASE =_ask_options( 'Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?' , ['Provide IAM Role name', 'Create new IAM role using credentials'] , _UpperCamelCase , ) if role_management == 0: _SCREAMING_SNAKE_CASE =_ask_field('Enter your IAM role name: ' ) else: _SCREAMING_SNAKE_CASE ='accelerate_sagemaker_execution_role' print(f"Accelerate will create an iam role \"{iam_role_name}\" using the provided credentials" ) _create_iam_role_for_sagemaker(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to use custom Docker image? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =None if is_custom_docker_image: _SCREAMING_SNAKE_CASE =_ask_field('Enter your Docker image: ' , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to provide SageMaker input channels with data locations? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =None if is_sagemaker_inputs_enabled: _SCREAMING_SNAKE_CASE =_ask_field( 'Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): ' , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to enable SageMaker metrics? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =None if is_sagemaker_metrics_enabled: _SCREAMING_SNAKE_CASE =_ask_field( 'Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): ' , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() , ) _SCREAMING_SNAKE_CASE =_ask_options( 'What is the distributed mode?' , ['No distributed training', 'Data parallelism'] , _convert_sagemaker_distributed_mode , ) _SCREAMING_SNAKE_CASE ={} _SCREAMING_SNAKE_CASE =_ask_field( 'Do you wish to optimize your script with torch dynamo?[yes/NO]:' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) if use_dynamo: _SCREAMING_SNAKE_CASE ='dynamo_' _SCREAMING_SNAKE_CASE =_ask_options( 'Which dynamo backend would you like to use?' , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to customize the defaults sent to torch.compile? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) if use_custom_options: _SCREAMING_SNAKE_CASE =_ask_options( 'Which mode do you want to use?' , _UpperCamelCase , lambda _UpperCamelCase : TORCH_DYNAMO_MODES[int(_UpperCamelCase )] , default='default' , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to enable dynamic shape tracing? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE ='Which EC2 instance type you want to use for your training?' if distributed_type != SageMakerDistributedType.NO: _SCREAMING_SNAKE_CASE =_ask_options( _UpperCamelCase , _UpperCamelCase , lambda _UpperCamelCase : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(_UpperCamelCase )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" _SCREAMING_SNAKE_CASE =_ask_field(_UpperCamelCase , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() , default='ml.p3.2xlarge' ) _SCREAMING_SNAKE_CASE =1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): _SCREAMING_SNAKE_CASE =_ask_field( 'How many machines do you want use? [1]: ' , _UpperCamelCase , default=1 , ) _SCREAMING_SNAKE_CASE =_ask_options( 'Do you wish to use FP16 or BF16 (mixed precision)?' , ['no', 'fp16', 'bf16', 'fp8'] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( 'Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.' ) return SageMakerConfig( image_uri=_UpperCamelCase , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=_UpperCamelCase , use_cpu=_UpperCamelCase , dynamo_config=_UpperCamelCase , eca_instance_type=_UpperCamelCase , profile=_UpperCamelCase , region=_UpperCamelCase , iam_role_name=_UpperCamelCase , mixed_precision=_UpperCamelCase , num_machines=_UpperCamelCase , sagemaker_inputs_file=_UpperCamelCase , sagemaker_metrics_file=_UpperCamelCase , )
47
1
import argparse from collections import defaultdict import yaml _UpperCAmelCase = """docs/source/en/_toctree.yml""" def UpperCamelCase ( __lowercase : List[str] ): '''simple docstring''' A_ : Optional[Any] = defaultdict(__lowercase ) A_ : int = [] A_ : Optional[Any] = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({'local': doc['local'], 'title': doc['title']} ) else: new_doc_list.append(__lowercase ) A_ : Tuple = new_doc_list A_ : Dict = [key for key, value in counts.items() if value > 1] A_ : int = [] for duplicate_key in duplicates: A_ : Optional[int] = list({doc['title'] for doc in doc_list if doc['local'] == duplicate_key} ) if len(__lowercase ) > 1: raise ValueError( f'''{duplicate_key} is present several times in the documentation table of content at ''' '`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ' 'others.' ) # Only add this once new_doc.append({'local': duplicate_key, 'title': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if 'local' not in counts or counts[doc['local']] == 1] ) A_ : str = sorted(__lowercase ,key=lambda __lowercase : s["title"].lower() ) # "overview" gets special treatment and is always first if len(__lowercase ) > 1: raise ValueError('{doc_list} has two \'overview\' docs which is not allowed.' ) overview_doc.extend(__lowercase ) # Sort return overview_doc def UpperCamelCase ( __lowercase : Dict=False ): '''simple docstring''' with open(__lowercase ,encoding='utf-8' ) as f: A_ : str = yaml.safe_load(f.read() ) # Get to the API doc A_ : Dict = 0 while content[api_idx]["title"] != "API": api_idx += 1 A_ : Dict = content[api_idx]['sections'] # Then to the model doc A_ : Optional[int] = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 A_ : int = api_doc[scheduler_idx]['sections'] A_ : Optional[Any] = clean_doc_toc(__lowercase ) A_ : Any = False if new_scheduler_doc != scheduler_doc: A_ : Tuple = True if overwrite: A_ : Tuple = new_scheduler_doc if diff: if overwrite: A_ : List[str] = api_doc with open(__lowercase ,'w' ,encoding='utf-8' ) as f: f.write(yaml.dump(__lowercase ,allow_unicode=__lowercase ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) def UpperCamelCase ( __lowercase : Dict=False ): '''simple docstring''' with open(__lowercase ,encoding='utf-8' ) as f: A_ : Any = yaml.safe_load(f.read() ) # Get to the API doc A_ : Dict = 0 while content[api_idx]["title"] != "API": api_idx += 1 A_ : Optional[Any] = content[api_idx]['sections'] # Then to the model doc A_ : Union[str, Any] = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 A_ : Optional[int] = False A_ : Union[str, Any] = api_doc[pipeline_idx]['sections'] A_ : Optional[Any] = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: A_ : List[str] = pipeline_doc['section'] A_ : Optional[int] = clean_doc_toc(__lowercase ) if overwrite: A_ : Optional[Any] = new_sub_pipeline_doc new_pipeline_docs.append(__lowercase ) # sort overall pipeline doc A_ : Tuple = clean_doc_toc(__lowercase ) if new_pipeline_docs != pipeline_docs: A_ : Tuple = True if overwrite: A_ : Any = new_pipeline_docs if diff: if overwrite: A_ : Optional[int] = api_doc with open(__lowercase ,'w' ,encoding='utf-8' ) as f: f.write(yaml.dump(__lowercase ,allow_unicode=__lowercase ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCAmelCase = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
362
from __future__ import annotations import requests _UpperCAmelCase = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def UpperCamelCase ( __lowercase : str ,__lowercase : int = 1 ,__lowercase : str = "new" ,__lowercase : list | None = None ): '''simple docstring''' A_ : Tuple = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(__lowercase ) - valid_terms ) ): A_ : int = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(__lowercase ) A_ : Optional[int] = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' ,headers={'User-agent': 'A random string'} ,) if response.status_code == 4_29: raise requests.HTTPError A_ : Optional[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(__lowercase )} A_ : Union[str, Any] = {} for id_ in range(__lowercase ): A_ : List[str] = { item: data['data']['children'][id_]['data'][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
192
0
"""simple docstring""" lowercase__ = [ 'DownloadConfig', 'DownloadManager', 'DownloadMode', 'StreamingDownloadManager', ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
290
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex lowercase__ = logging.getLogger(__name__) class __snake_case : def __init__( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' if not self.initialized: a__: Optional[int] = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Optional[int] = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.retriever.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ , a__: str = self.retriever._main_retrieve(lowercase , lowercase) return doc_ids, retrieved_doc_embeds class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> int: '''simple docstring''' if index is not None and index.is_initialized() and len(lowercase) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ') super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Any = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase) for worker in self.retrieval_workers ]) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' logger.info('initializing retrieval') if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if len(self.retrieval_workers) > 0: # Select a random retrieval actor. a__: int = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] a__ , a__: List[Any] = ray.get(random_worker.retrieve.remote(lowercase , lowercase)) else: a__ , a__: Dict = self._main_retrieve(lowercase , lowercase) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' return super(lowercase , cls).get_tokenizers(lowercase , lowercase , **lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = kwargs.pop('config' , lowercase) or RagConfig.from_pretrained(lowercase , **lowercase) a__: Union[str, Any] = RagTokenizer.from_pretrained(lowercase , config=lowercase) a__: int = rag_tokenizer.question_encoder a__: Any = rag_tokenizer.generator if indexed_dataset is not None: a__: List[Any] = 'custom' a__: Optional[Any] = CustomHFIndex(config.retrieval_vector_size , lowercase) else: a__: Dict = cls._build_index(lowercase) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
290
1
def lowerCAmelCase_ ( __a ) -> Optional[int]: """simple docstring""" if collection == []: return [] # get some information about the collection lowerCamelCase__: int =len(__a ) lowerCamelCase__: str =max(__a ) lowerCamelCase__: Optional[Any] =min(__a ) # create the counting array lowerCamelCase__: List[str] =coll_max + 1 - coll_min lowerCamelCase__: Optional[Any] =[0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , __a ): lowerCamelCase__: List[Any] =counting_arr[i] + counting_arr[i - 1] # create the output collection lowerCamelCase__: List[str] =[0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , __a ) ): lowerCamelCase__: Tuple =collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def lowerCAmelCase_ ( __a ) -> str: """simple docstring""" return "".join([chr(__a ) for i in counting_sort([ord(__a ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("thisisthestring") == "eghhiiinrsssttt" __A = input("Enter numbers separated by a comma:\n").strip() __A = [int(item) for item in user_input.split(",")] print(counting_sort(unsorted))
357
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __A = "src/transformers" # This is to make sure the transformers module imported is the one in the repo. __A = direct_transformers_import(PATH_TO_TRANSFORMERS) __A = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __A = re.compile(R"\[(.+?)\]\((https://huggingface\.co/.+?)\)") __A = { "DecisionTransformerConfig", "EncoderDecoderConfig", "MusicgenConfig", "RagConfig", "SpeechEncoderDecoderConfig", "TimmBackboneConfig", "VisionEncoderDecoderConfig", "VisionTextDualEncoderConfig", "LlamaConfig", } def lowerCAmelCase_ ( __a ) -> Optional[int]: """simple docstring""" lowerCamelCase__: List[Any] =None # source code of `config_class` lowerCamelCase__: List[Any] =inspect.getsource(__a ) lowerCamelCase__: str =_re_checkpoint.findall(__a ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("/" ): lowerCamelCase__: str =ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCamelCase__: Optional[int] =F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: lowerCamelCase__: List[Any] =ckpt_name break return checkpoint def lowerCAmelCase_ ( ) -> Tuple: """simple docstring""" lowerCamelCase__: Union[str, Any] =[] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCamelCase__: List[Any] =get_checkpoint_from_config_class(__a ) lowerCamelCase__: Tuple =config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__a ) if len(__a ) > 0: lowerCamelCase__: List[str] ="\n".join(sorted(__a ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
273
0
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class _a (A__ , A__ , A__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: List[Any] = StableUnCLIPPipeline UpperCAmelCase__: int = TEXT_TO_IMAGE_PARAMS UpperCAmelCase__: Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS UpperCAmelCase__: Union[str, Any] = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase__: int = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false UpperCAmelCase__: List[str] = False def __A ( self ): A__ : List[str] = 32 A__ : List[Any] = embedder_hidden_size # prior components torch.manual_seed(0 ) A__ : Dict = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) torch.manual_seed(0 ) A__ : Any = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=A__ , projection_dim=A__ , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) A__ : str = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=A__ , num_layers=1 , ) torch.manual_seed(0 ) A__ : List[Any] = DDPMScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1000 , clip_sample=A__ , clip_sample_range=5.0 , beta_schedule="""squaredcos_cap_v2""" , ) # regular denoising components torch.manual_seed(0 ) A__ : List[str] = StableUnCLIPImageNormalizer(embedding_dim=A__ ) A__ : Optional[int] = DDPMScheduler(beta_schedule="""squaredcos_cap_v2""" ) torch.manual_seed(0 ) A__ : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) torch.manual_seed(0 ) A__ : Any = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=A__ , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) A__ : Any = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """CrossAttnUpBlock2D""") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="""projection""" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=A__ , layers_per_block=1 , upcast_attention=A__ , use_linear_projection=A__ , ) torch.manual_seed(0 ) A__ : Optional[Any] = DDIMScheduler( beta_schedule="""scaled_linear""" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , prediction_type="""v_prediction""" , set_alpha_to_one=A__ , steps_offset=1 , ) torch.manual_seed(0 ) A__ : List[Any] = AutoencoderKL() A__ : str = { # prior components """prior_tokenizer""": prior_tokenizer, """prior_text_encoder""": prior_text_encoder, """prior""": prior, """prior_scheduler""": prior_scheduler, # image noising components """image_normalizer""": image_normalizer, """image_noising_scheduler""": image_noising_scheduler, # regular denoising components """tokenizer""": tokenizer, """text_encoder""": text_encoder, """unet""": unet, """scheduler""": scheduler, """vae""": vae, } return components def __A ( self , A__ , A__=0 ): if str(A__ ).startswith("""mps""" ): A__ : str = torch.manual_seed(A__ ) else: A__ : Optional[Any] = torch.Generator(device=A__ ).manual_seed(A__ ) A__ : Tuple = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """prior_num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def __A ( self ): A__ : Dict = torch_device == """cpu""" self._test_attention_slicing_forward_pass(test_max_difference=A__ ) def __A ( self ): A__ : Union[str, Any] = torch_device in ["""cpu""", """mps"""] self._test_inference_batch_single_identical(test_max_difference=A__ ) @slow @require_torch_gpu class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self ): A__ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy""" ) A__ : str = StableUnCLIPPipeline.from_pretrained("""fusing/stable-unclip-2-1-l""" , torch_dtype=torch.floataa ) pipe.to(A__ ) pipe.set_progress_bar_config(disable=A__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() A__ : Tuple = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : Dict = pipe("""anime turle""" , generator=A__ , output_type="""np""" ) A__ : List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(A__ , A__ ) def __A ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() A__ : Tuple = StableUnCLIPPipeline.from_pretrained("""fusing/stable-unclip-2-1-l""" , torch_dtype=torch.floataa ) A__ : int = pipe.to(A__ ) pipe.set_progress_bar_config(disable=A__ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() A__ : Dict = pipe( """anime turtle""" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="""np""" , ) A__ : Optional[int] = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
192
import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() a_ : str = logging.get_logger(__name__) def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase): SCREAMING_SNAKE_CASE = RobertaPreLayerNormConfig.from_pretrained( _UpperCAmelCase , architectures=['RobertaPreLayerNormForMaskedLM']) # convert state_dict SCREAMING_SNAKE_CASE = torch.load(hf_hub_download(repo_id=_UpperCAmelCase , filename='pytorch_model.bin')) SCREAMING_SNAKE_CASE = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith('roberta.'): SCREAMING_SNAKE_CASE = 'roberta_prelayernorm.' + tensor_key[len('roberta.') :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith('.self.LayerNorm.weight') or tensor_key.endswith('.self.LayerNorm.bias'): continue SCREAMING_SNAKE_CASE = tensor_value SCREAMING_SNAKE_CASE = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=_UpperCAmelCase , config=_UpperCAmelCase , state_dict=_UpperCAmelCase) model.save_pretrained(_UpperCAmelCase) # convert tokenizer SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(_UpperCAmelCase) tokenizer.save_pretrained(_UpperCAmelCase) if __name__ == "__main__": a_ : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint-repo', default=None, type=str, required=True, help='Path the official PyTorch dump, e.g. \'andreasmadsen/efficient_mlm_m0.40\'.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) a_ : int = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
137
0
"""simple docstring""" import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class _A ( lowerCAmelCase ): def __init__( self , __lowerCAmelCase ): """simple docstring""" lowercase = data def __iter__( self ): """simple docstring""" for element in self.data: yield element def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[Any]=True ) -> List[Any]: '''simple docstring''' lowercase = Accelerator(even_batches=lowerCAmelCase__ ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def UpperCAmelCase__ ( lowerCAmelCase__ :Accelerator , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :bool = False ) -> Dict: '''simple docstring''' if iterable: lowercase = DummyIterableDataset(torch.as_tensor(range(lowerCAmelCase__ ) ) ) else: lowercase = TensorDataset(torch.as_tensor(range(lowerCAmelCase__ ) ) ) lowercase = DataLoader(lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) lowercase = accelerator.prepare(lowerCAmelCase__ ) return dl def UpperCAmelCase__ ( lowerCAmelCase__ :Accelerator , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :List[int] , ) -> Dict: '''simple docstring''' lowercase = create_dataloader(accelerator=lowerCAmelCase__ , dataset_size=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) lowercase = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def UpperCAmelCase__ ( ) -> Any: '''simple docstring''' lowercase = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( lowerCAmelCase__ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( lowerCAmelCase__ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def UpperCAmelCase__ ( ) -> Optional[Any]: '''simple docstring''' lowercase = create_accelerator(even_batches=lowerCAmelCase__ ) verify_dataloader_batch_sizes( lowerCAmelCase__ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( lowerCAmelCase__ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def UpperCAmelCase__ ( ) -> Dict: '''simple docstring''' lowercase = create_accelerator(even_batches=lowerCAmelCase__ ) lowercase = torch.nn.Linear(1 , 1 ) lowercase = accelerator.prepare(lowerCAmelCase__ ) lowercase = create_dataloader(lowerCAmelCase__ , dataset_size=3 , batch_size=1 ) lowercase = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(lowerCAmelCase__ ): lowercase = ddp_model(batch[0].float() ) lowercase = output.sum() loss.backward() batch_idxs.append(lowerCAmelCase__ ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] ) -> Tuple: '''simple docstring''' with warnings.catch_warnings(record=lowerCAmelCase__ ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , lowerCAmelCase__ ) assert "only supported for multi-GPU" in str(w[-1].message ) def UpperCAmelCase__ ( ) -> Optional[int]: '''simple docstring''' lowercase = True lowercase = False lowercase = create_accelerator(even_batches=lowerCAmelCase__ ) lowercase = torch.nn.Linear(1 , 1 ) lowercase = accelerator.prepare(lowerCAmelCase__ ) lowercase = create_dataloader(lowerCAmelCase__ , dataset_size=3 , batch_size=1 ) lowercase = create_dataloader(lowerCAmelCase__ , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase__ ): lowercase = train_dl.batch_sampler.even_batches lowercase = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def UpperCAmelCase__ ( ) -> int: '''simple docstring''' lowercase = True lowercase = False lowercase = create_accelerator(even_batches=lowerCAmelCase__ ) lowercase = torch.nn.Linear(1 , 1 ) lowercase = accelerator.prepare(lowerCAmelCase__ ) create_dataloader(lowerCAmelCase__ , dataset_size=3 , batch_size=1 , iterable=lowerCAmelCase__ ) lowercase = create_dataloader(lowerCAmelCase__ , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings("""ignore""" ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase__ ): lowercase = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def UpperCAmelCase__ ( ) -> Optional[int]: '''simple docstring''' lowercase = create_accelerator() lowercase = torch.nn.Linear(1 , 1 ) lowercase = accelerator.prepare(lowerCAmelCase__ ) create_dataloader(lowerCAmelCase__ , dataset_size=3 , batch_size=1 , iterable=lowerCAmelCase__ ) with warnings.catch_warnings(record=lowerCAmelCase__ ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase__ ): pass assert issubclass(w[-1].category , lowerCAmelCase__ ) assert "only supported for map-style datasets" in str(w[-1].message ) def UpperCAmelCase__ ( ) -> Tuple: '''simple docstring''' lowercase = create_accelerator() accelerator.print("""Test that even_batches variable ensures uniform batches across processes""" ) test_default_ensures_even_batch_sizes() accelerator.print("""Run tests with even_batches disabled""" ) test_can_disable_even_batches() accelerator.print("""Test joining uneven inputs""" ) test_can_join_uneven_inputs() accelerator.print("""Test overriding even_batches when joining uneven inputs""" ) test_join_can_override_even_batches() accelerator.print("""Test overriding even_batches for mixed dataloader types""" ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print("""Test overriding even_batches raises a warning for iterable dataloaders""" ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print("""Test join with non DDP distributed raises warning""" ) lowercase = accelerator.state.distributed_type lowercase = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(lowerCAmelCase__ ) lowercase = original_state if __name__ == "__main__": main()
368
"""simple docstring""" def UpperCAmelCase__ ( lowerCAmelCase__ :str ) -> bool: '''simple docstring''' return credit_card_number.startswith(("""34""", """35""", """37""", """4""", """5""", """6""") ) def UpperCAmelCase__ ( lowerCAmelCase__ :str ) -> bool: '''simple docstring''' lowercase = credit_card_number lowercase = 0 lowercase = len(lowerCAmelCase__ ) - 2 for i in range(lowerCAmelCase__ , -1 , -2 ): # double the value of every second digit lowercase = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 1_0 digit += 1 lowercase = cc_number[:i] + str(lowerCAmelCase__ ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(lowerCAmelCase__ ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 1_0 == 0 def UpperCAmelCase__ ( lowerCAmelCase__ :str ) -> bool: '''simple docstring''' lowercase = f'{credit_card_number} is an invalid credit card number because' if not credit_card_number.isdigit(): print(f'{error_message} it has nonnumerical characters.' ) return False if not 1_3 <= len(lowerCAmelCase__ ) <= 1_6: print(f'{error_message} of its length.' ) return False if not validate_initial_digits(lowerCAmelCase__ ): print(f'{error_message} of its first two digits.' ) return False if not luhn_validation(lowerCAmelCase__ ): print(f'{error_message} it fails the Luhn check.' ) return False print(f'{credit_card_number} is a valid credit card number.' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number("""4111111111111111""") validate_credit_card_number("""32323""")
32
0
"""simple docstring""" def A ( snake_case :list[int] ) -> int: if not numbers: return 0 if not isinstance(snake_case , (list, tuple) ) or not all( isinstance(snake_case , snake_case ) for number in numbers ): raise ValueError('numbers must be an iterable of integers' ) __UpperCamelCase = __UpperCamelCase = __UpperCamelCase = numbers[0] for i in range(1 , len(snake_case ) ): # update the maximum and minimum subarray products __UpperCamelCase = numbers[i] if number < 0: __UpperCamelCase , __UpperCamelCase = min_till_now, max_till_now __UpperCamelCase = max(snake_case , max_till_now * number ) __UpperCamelCase = min(snake_case , min_till_now * number ) # update the maximum product found till now __UpperCamelCase = max(snake_case , snake_case ) return max_prod
316
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging UpperCamelCase : Any = logging.get_logger(__name__) class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = ["pixel_values"] def __init__( self , __UpperCAmelCase = True , __UpperCAmelCase = 1 / 255 , __UpperCAmelCase = True , __UpperCAmelCase = 8 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) __UpperCamelCase = do_rescale __UpperCamelCase = rescale_factor __UpperCamelCase = do_pad __UpperCamelCase = pad_size def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase ): '''simple docstring''' return rescale(__UpperCAmelCase , scale=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase = get_image_size(__UpperCAmelCase ) __UpperCamelCase = (old_height // size + 1) * size - old_height __UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__UpperCAmelCase , ((0, pad_height), (0, pad_width)) , mode='symmetric' , data_format=__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = ChannelDimension.FIRST , **__UpperCAmelCase , ): '''simple docstring''' __UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCamelCase = do_pad if do_pad is not None else self.do_pad __UpperCamelCase = pad_size if pad_size is not None else self.pad_size __UpperCamelCase = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) # All transformations expect numpy arrays. __UpperCamelCase = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_rescale: __UpperCamelCase = [self.rescale(image=__UpperCAmelCase , scale=__UpperCAmelCase ) for image in images] if do_pad: __UpperCamelCase = [self.pad(__UpperCAmelCase , size=__UpperCAmelCase ) for image in images] __UpperCamelCase = [to_channel_dimension_format(__UpperCAmelCase , __UpperCAmelCase ) for image in images] __UpperCamelCase = {'pixel_values': images} return BatchFeature(data=__UpperCAmelCase , tensor_type=__UpperCAmelCase )
316
1
def lowercase_ ( _lowerCamelCase : Union[str, Any]): lowercase__ : List[Any] = [] lowercase__ : Any = set({"(", "[", "{"}) lowercase__ : Any = set({")", "]", "}"}) lowercase__ : Any = {"{": "}", "[": "]", "(": ")"} for i in range(len(_lowerCamelCase)): if s[i] in open_brackets: stack.append(s[i]) elif s[i] in closed_brackets and ( len(_lowerCamelCase) == 0 or (len(_lowerCamelCase) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(_lowerCamelCase) == 0 def lowercase_ ( ): lowercase__ : int = input("Enter sequence of brackets: ") if is_balanced(_lowerCamelCase): print(_lowerCamelCase , "is balanced") else: print(_lowerCamelCase , "is not balanced") if __name__ == "__main__": main()
333
def lowercase_ ( _lowerCamelCase : int , _lowerCamelCase : int): while a != 0: lowercase__ , lowercase__ : Dict = b % a, a return b def lowercase_ ( _lowerCamelCase : int , _lowerCamelCase : int): if gcd(_lowerCamelCase , _lowerCamelCase) != 1: lowercase__ : Tuple = f'''mod inverse of {a!r} and {m!r} does not exist''' raise ValueError(_lowerCamelCase) lowercase__ , lowercase__ , lowercase__ : Optional[int] = 1, 0, a lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = 0, 1, m while va != 0: lowercase__ : Tuple = ua // va lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : Any = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
333
1
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = '▁' UpperCamelCase_ = { 'vocab_file': 'vocab.json', 'spm_file': 'sentencepiece.bpe.model', 'tokenizer_config_file': 'tokenizer_config.json', } UpperCamelCase_ = { 'vocab_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json', }, 'spm_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_config_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json', }, } UpperCamelCase_ = { 'facebook/m2m100_418M': 1024, } # fmt: off UpperCamelCase_ = { 'm2m100': ['af', 'am', 'ar', 'ast', 'az', 'ba', 'be', 'bg', 'bn', 'br', 'bs', 'ca', 'ceb', 'cs', 'cy', 'da', 'de', 'el', 'en', 'es', 'et', 'fa', 'ff', 'fi', 'fr', 'fy', 'ga', 'gd', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hy', 'id', 'ig', 'ilo', 'is', 'it', 'ja', 'jv', 'ka', 'kk', 'km', 'kn', 'ko', 'lb', 'lg', 'ln', 'lo', 'lt', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'my', 'ne', 'nl', 'no', 'ns', 'oc', 'or', 'pa', 'pl', 'ps', 'pt', 'ro', 'ru', 'sd', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'ss', 'su', 'sv', 'sw', 'ta', 'th', 'tl', 'tn', 'tr', 'uk', 'ur', 'uz', 'vi', 'wo', 'xh', 'yi', 'yo', 'zh', 'zu'], 'wmt21': ['en', 'ha', 'is', 'ja', 'cs', 'ru', 'zh', 'de'] } class snake_case ( SCREAMING_SNAKE_CASE_ ): a_ : List[str] = VOCAB_FILES_NAMES a_ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ : Any = PRETRAINED_VOCAB_FILES_MAP a_ : List[Any] = ["""input_ids""", """attention_mask"""] a_ : List[int] = [] a_ : List[int] = [] def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase="<s>" , __UpperCAmelCase="</s>" , __UpperCAmelCase="</s>" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="m2m100" , __UpperCAmelCase = None , __UpperCAmelCase=8 , **__UpperCAmelCase , ) ->None: a_ = {} if sp_model_kwargs is None else sp_model_kwargs a_ = language_codes a_ = FAIRSEQ_LANGUAGE_CODES[language_codes] a_ = {lang_code: F'''__{lang_code}__''' for lang_code in fairseq_language_code} a_ = kwargs.get("additional_special_tokens" , []) kwargs["additional_special_tokens"] += [ self.get_lang_token(__UpperCAmelCase) for lang_code in fairseq_language_code if self.get_lang_token(__UpperCAmelCase) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__UpperCAmelCase , tgt_lang=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , language_codes=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=__UpperCAmelCase , **__UpperCAmelCase , ) a_ = vocab_file a_ = load_json(__UpperCAmelCase) a_ = {v: k for k, v in self.encoder.items()} a_ = spm_file a_ = load_spm(__UpperCAmelCase , self.sp_model_kwargs) a_ = len(self.encoder) a_ = { self.get_lang_token(__UpperCAmelCase): self.encoder_size + i for i, lang_code in enumerate(__UpperCAmelCase) } a_ = {lang_code: self.encoder_size + i for i, lang_code in enumerate(__UpperCAmelCase)} a_ = {v: k for k, v in self.lang_token_to_id.items()} a_ = src_lang if src_lang is not None else "en" a_ = tgt_lang a_ = self.get_lang_id(self._src_lang) self.set_src_lang_special_tokens(self._src_lang) a_ = num_madeup_words @property def UpperCAmelCase__ ( self) ->int: return len(self.encoder) + len(self.lang_token_to_id) @property def UpperCAmelCase__ ( self) ->str: return self._src_lang @src_lang.setter def UpperCAmelCase__ ( self , __UpperCAmelCase) ->None: a_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->List[str]: return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->str: if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(__UpperCAmelCase , self.encoder[self.unk_token]) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->str: if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(__UpperCAmelCase , self.unk_token) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->Union[str, Any]: a_ = [] a_ = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__UpperCAmelCase) + token a_ = [] else: current_sub_tokens.append(__UpperCAmelCase) out_string += self.sp_model.decode(__UpperCAmelCase) return out_string.strip() def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False) ->List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase) a_ = [1] * len(self.prefix_tokens) a_ = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(__UpperCAmelCase)) + suffix_ones return prefix_ones + ([0] * len(__UpperCAmelCase)) + ([0] * len(__UpperCAmelCase)) + suffix_ones def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCAmelCase__ ( self) ->Dict: a_ = {self.convert_ids_to_tokens(__UpperCAmelCase): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) ->Dict: a_ = self.__dict__.copy() a_ = None return state def __setstate__( self , __UpperCAmelCase) ->None: a_ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): a_ = {} a_ = load_spm(self.spm_file , self.sp_model_kwargs) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->Tuple[str]: a_ = Path(__UpperCAmelCase) if not save_dir.is_dir(): raise OSError(F'''{save_directory} should be a directory''') a_ = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) a_ = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , __UpperCAmelCase) if os.path.abspath(self.spm_file) != os.path.abspath(__UpperCAmelCase) and os.path.isfile(self.spm_file): copyfile(self.spm_file , __UpperCAmelCase) elif not os.path.isfile(self.spm_file): with open(__UpperCAmelCase , "wb") as fi: a_ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase) return (str(__UpperCAmelCase), str(__UpperCAmelCase)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = "en" , __UpperCAmelCase = None , __UpperCAmelCase = "ro" , **__UpperCAmelCase , ) ->BatchEncoding: a_ = src_lang a_ = tgt_lang self.set_src_lang_special_tokens(self.src_lang) return super().prepare_seqaseq_batch(__UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase) ->List[Any]: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model") a_ = src_lang a_ = self(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , **__UpperCAmelCase) a_ = self.get_lang_id(__UpperCAmelCase) a_ = tgt_lang_id return inputs def UpperCAmelCase__ ( self) ->Any: self.set_src_lang_special_tokens(self.src_lang) def UpperCAmelCase__ ( self) ->str: self.set_tgt_lang_special_tokens(self.tgt_lang) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->None: a_ = self.get_lang_token(__UpperCAmelCase) a_ = self.lang_token_to_id[lang_token] a_ = [self.cur_lang_id] a_ = [self.eos_token_id] def UpperCAmelCase__ ( self , __UpperCAmelCase) ->None: a_ = self.get_lang_token(__UpperCAmelCase) a_ = self.lang_token_to_id[lang_token] a_ = [self.cur_lang_id] a_ = [self.eos_token_id] def UpperCAmelCase__ ( self , __UpperCAmelCase) ->str: return self.lang_code_to_token[lang] def UpperCAmelCase__ ( self , __UpperCAmelCase) ->int: a_ = self.get_lang_token(__UpperCAmelCase) return self.lang_token_to_id[lang_token] def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ) ->sentencepiece.SentencePieceProcessor: """simple docstring""" a_ = sentencepiece.SentencePieceProcessor(**UpperCAmelCase ) spm.Load(str(UpperCAmelCase ) ) return spm def UpperCamelCase ( UpperCAmelCase ) ->Union[Dict, List]: """simple docstring""" with open(UpperCAmelCase , "r" ) as f: return json.load(UpperCAmelCase ) def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ) ->None: """simple docstring""" with open(UpperCAmelCase , "w" ) as f: json.dump(UpperCAmelCase , UpperCAmelCase , indent=2 )
243
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCamelCase_ = 'src/diffusers' UpperCamelCase_ = '.' # This is to make sure the diffusers module imported is the one in the repo. UpperCamelCase_ = importlib.util.spec_from_file_location( 'diffusers', os.path.join(DIFFUSERS_PATH, '__init__.py'), submodule_search_locations=[DIFFUSERS_PATH], ) UpperCamelCase_ = spec.loader.load_module() def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ) ->Dict: """simple docstring""" return line.startswith(UpperCAmelCase ) or len(UpperCAmelCase ) <= 1 or re.search(r"^\s*\)(\s*->.*:|:)\s*$" , UpperCAmelCase ) is not None def UpperCamelCase ( UpperCAmelCase ) ->Any: """simple docstring""" a_ = object_name.split("." ) a_ = 0 # First let's find the module where our object lives. a_ = parts[i] while i < len(UpperCAmelCase ) and not os.path.isfile(os.path.join(UpperCAmelCase , F'''{module}.py''' ) ): i += 1 if i < len(UpperCAmelCase ): a_ = os.path.join(UpperCAmelCase , parts[i] ) if i >= len(UpperCAmelCase ): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''' ) with open(os.path.join(UpperCAmelCase , F'''{module}.py''' ) , "r" , encoding="utf-8" , newline="\n" ) as f: a_ = f.readlines() # Now let's find the class / func in the code! a_ = "" a_ = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase ) and re.search(rF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase ): raise ValueError(F''' {object_name} does not match any function or class in {module}.''' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). a_ = line_index while line_index < len(UpperCAmelCase ) and _should_continue(lines[line_index] , UpperCAmelCase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 a_ = lines[start_index:line_index] return "".join(UpperCAmelCase ) UpperCamelCase_ = re.compile(R'^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)') UpperCamelCase_ = re.compile(R'^\s*(\S+)->(\S+)(\s+.*|$)') UpperCamelCase_ = re.compile(R'<FILL\s+[^>]*>') def UpperCamelCase ( UpperCAmelCase ) ->int: """simple docstring""" a_ = code.split("\n" ) a_ = 0 while idx < len(UpperCAmelCase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase ): return re.search(r"^(\s*)\S" , lines[idx] ).groups()[0] return "" def UpperCamelCase ( UpperCAmelCase ) ->int: """simple docstring""" a_ = len(get_indent(UpperCAmelCase ) ) > 0 if has_indent: a_ = F'''class Bla:\n{code}''' a_ = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=UpperCAmelCase ) a_ = black.format_str(UpperCAmelCase , mode=UpperCAmelCase ) a_ , a_ = style_docstrings_in_code(UpperCAmelCase ) return result[len("class Bla:\n" ) :] if has_indent else result def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase=False ) ->str: """simple docstring""" with open(UpperCAmelCase , "r" , encoding="utf-8" , newline="\n" ) as f: a_ = f.readlines() a_ = [] a_ = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase ): a_ = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. a_ , a_ , a_ = search.groups() a_ = find_code_in_diffusers(UpperCAmelCase ) a_ = get_indent(UpperCAmelCase ) a_ = line_index + 1 if indent == theoretical_indent else line_index + 2 a_ = theoretical_indent a_ = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. a_ = True while line_index < len(UpperCAmelCase ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase ): break a_ = lines[line_index] a_ = _should_continue(UpperCAmelCase , UpperCAmelCase ) and re.search(F'''^{indent}# End copy''' , UpperCAmelCase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 a_ = lines[start_index:line_index] a_ = "".join(UpperCAmelCase ) # Remove any nested `Copied from` comments to avoid circular copies a_ = [line for line in theoretical_code.split("\n" ) if _re_copy_warning.search(UpperCAmelCase ) is None] a_ = "\n".join(UpperCAmelCase ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase ) > 0: a_ = replace_pattern.replace("with" , "" ).split("," ) a_ = [_re_replace_pattern.search(UpperCAmelCase ) for p in patterns] for pattern in patterns: if pattern is None: continue a_ , a_ , a_ = pattern.groups() a_ = re.sub(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if option.strip() == "all-casing": a_ = re.sub(obja.lower() , obja.lower() , UpperCAmelCase ) a_ = re.sub(obja.upper() , obja.upper() , UpperCAmelCase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line a_ = blackify(lines[start_index - 1] + theoretical_code ) a_ = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: a_ = lines[:start_index] + [theoretical_code] + lines[line_index:] a_ = start_index + 1 if overwrite and len(UpperCAmelCase ) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''' ) with open(UpperCAmelCase , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(UpperCAmelCase ) return diffs def UpperCamelCase ( UpperCAmelCase = False ) ->int: """simple docstring""" a_ = glob.glob(os.path.join(UpperCAmelCase , "**/*.py" ) , recursive=UpperCAmelCase ) a_ = [] for filename in all_files: a_ = is_copy_consistent(UpperCAmelCase , UpperCAmelCase ) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(UpperCAmelCase ) > 0: a_ = "\n".join(UpperCAmelCase ) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them." ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') UpperCamelCase_ = parser.parse_args() check_copies(args.fix_and_overwrite)
243
1
import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 snake_case : str = get_tests_dir('''fixtures''') class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): # A mock response for an HTTP head request to emulate server down a :List[str] = mock.Mock() a :Dict = 500 a :Any = {} a :Optional[int] = HTTPError a :Union[str, Any] = {} # Download this model to make sure it's in the cache. a :Union[str, Any] = ViTImageProcessor.from_pretrained('''hf-internal-testing/tiny-random-vit''' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('''requests.Session.request''' , return_value=_lowerCamelCase ) as mock_head: a :Optional[int] = ViTImageProcessor.from_pretrained('''hf-internal-testing/tiny-random-vit''' ) # This check we did call the fake head request mock_head.assert_called() def SCREAMING_SNAKE_CASE__ ( self ): # This test is for deprecated behavior and can be removed in v5 a :Optional[int] = ViTImageProcessor.from_pretrained( '''https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json''' ) def SCREAMING_SNAKE_CASE__ ( self ): with self.assertRaises(_lowerCamelCase ): # config is in subfolder, the following should not work without specifying the subfolder a :str = AutoImageProcessor.from_pretrained('''hf-internal-testing/stable-diffusion-all-variants''' ) a :Dict = AutoImageProcessor.from_pretrained( '''hf-internal-testing/stable-diffusion-all-variants''' , subfolder='''feature_extractor''' ) self.assertIsNotNone(_lowerCamelCase ) @is_staging_test class _snake_case ( unittest.TestCase ): @classmethod def SCREAMING_SNAKE_CASE__ ( cls ): a :Optional[Any] = TOKEN HfFolder.save_token(_lowerCamelCase ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls ): try: delete_repo(token=cls._token , repo_id='''test-image-processor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-image-processor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-image-processor''' ) except HTTPError: pass def SCREAMING_SNAKE_CASE__ ( self ): a :List[Any] = ViTImageProcessor.from_pretrained(_lowerCamelCase ) image_processor.push_to_hub('''test-image-processor''' , use_auth_token=self._token ) a :Any = ViTImageProcessor.from_pretrained(F'''{USER}/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id='''test-image-processor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( _lowerCamelCase , repo_id='''test-image-processor''' , push_to_hub=_lowerCamelCase , use_auth_token=self._token ) a :Optional[Any] = ViTImageProcessor.from_pretrained(F'''{USER}/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self ): a :Dict = ViTImageProcessor.from_pretrained(_lowerCamelCase ) image_processor.push_to_hub('''valid_org/test-image-processor''' , use_auth_token=self._token ) a :str = ViTImageProcessor.from_pretrained('''valid_org/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-image-processor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( _lowerCamelCase , repo_id='''valid_org/test-image-processor-org''' , push_to_hub=_lowerCamelCase , use_auth_token=self._token ) a :Union[str, Any] = ViTImageProcessor.from_pretrained('''valid_org/test-image-processor-org''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self ): CustomImageProcessor.register_for_auto_class() a :Optional[int] = CustomImageProcessor.from_pretrained(_lowerCamelCase ) image_processor.push_to_hub('''test-dynamic-image-processor''' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map , {'''AutoImageProcessor''': '''custom_image_processing.CustomImageProcessor'''} , ) a :str = AutoImageProcessor.from_pretrained( F'''{USER}/test-dynamic-image-processor''' , trust_remote_code=_lowerCamelCase ) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__ , '''CustomImageProcessor''' )
281
def __lowerCamelCase ( UpperCAmelCase_ : str ): """simple docstring""" if n_term == "": return [] a :list = [] for temp in range(int(UpperCAmelCase_ ) ): series.append(F'''1/{temp + 1}''' if series else '''1''' ) return series if __name__ == "__main__": snake_case : Tuple = input('''Enter the last number (nth term) of the Harmonic Series''') print('''Formula of Harmonic Series => 1+1/2+1/3 ..... 1/n''') print(harmonic_series(nth_term))
281
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Optional[Any] =logging.get_logger(__name__) _UpperCAmelCase : int ={ 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class snake_case__( __UpperCAmelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = """unispeech-sat""" def __init__( self , __lowercase=3_2 , __lowercase=7_6_8 , __lowercase=1_2 , __lowercase=1_2 , __lowercase=3_0_7_2 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.02 , __lowercase=1e-5 , __lowercase="group" , __lowercase="gelu" , __lowercase=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2) , __lowercase=(5, 2, 2, 2, 2, 2, 2) , __lowercase=(1_0, 3, 3, 3, 3, 2, 2) , __lowercase=False , __lowercase=1_2_8 , __lowercase=1_6 , __lowercase=False , __lowercase=True , __lowercase=0.05 , __lowercase=1_0 , __lowercase=2 , __lowercase=0.0 , __lowercase=1_0 , __lowercase=0 , __lowercase=3_2_0 , __lowercase=2 , __lowercase=0.1 , __lowercase=1_0_0 , __lowercase=2_5_6 , __lowercase=2_5_6 , __lowercase=0.1 , __lowercase="mean" , __lowercase=False , __lowercase=False , __lowercase=2_5_6 , __lowercase=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 1_5_0_0) , __lowercase=(5, 3, 3, 1, 1) , __lowercase=(1, 2, 3, 1, 1) , __lowercase=5_1_2 , __lowercase=0 , __lowercase=1 , __lowercase=2 , __lowercase=5_0_4 , **__lowercase , ) -> List[Any]: super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase ) lowerCAmelCase_ : Any = hidden_size lowerCAmelCase_ : Optional[Any] = feat_extract_norm lowerCAmelCase_ : str = feat_extract_activation lowerCAmelCase_ : Optional[Any] = list(_lowerCamelCase ) lowerCAmelCase_ : Tuple = list(_lowerCamelCase ) lowerCAmelCase_ : Dict = list(_lowerCamelCase ) lowerCAmelCase_ : int = conv_bias lowerCAmelCase_ : Tuple = num_conv_pos_embeddings lowerCAmelCase_ : Dict = num_conv_pos_embedding_groups lowerCAmelCase_ : Optional[Any] = len(self.conv_dim ) lowerCAmelCase_ : Optional[int] = num_hidden_layers lowerCAmelCase_ : Optional[int] = intermediate_size lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : List[str] = num_attention_heads lowerCAmelCase_ : Dict = hidden_dropout lowerCAmelCase_ : Union[str, Any] = attention_dropout lowerCAmelCase_ : Dict = activation_dropout lowerCAmelCase_ : str = feat_proj_dropout lowerCAmelCase_ : Optional[int] = final_dropout lowerCAmelCase_ : str = layerdrop lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : str = vocab_size lowerCAmelCase_ : Tuple = num_clusters lowerCAmelCase_ : Dict = do_stable_layer_norm lowerCAmelCase_ : Any = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowerCAmelCase_ : Optional[int] = apply_spec_augment lowerCAmelCase_ : Dict = mask_time_prob lowerCAmelCase_ : List[Any] = mask_time_length lowerCAmelCase_ : str = mask_time_min_masks lowerCAmelCase_ : List[str] = mask_feature_prob lowerCAmelCase_ : int = mask_feature_length lowerCAmelCase_ : Any = mask_feature_min_masks # parameters for pretraining with codevector quantized representations lowerCAmelCase_ : int = num_codevectors_per_group lowerCAmelCase_ : str = num_codevector_groups lowerCAmelCase_ : List[str] = contrastive_logits_temperature lowerCAmelCase_ : Tuple = feat_quantizer_dropout lowerCAmelCase_ : Dict = num_negatives lowerCAmelCase_ : Union[str, Any] = codevector_dim lowerCAmelCase_ : Optional[int] = proj_codevector_dim lowerCAmelCase_ : Union[str, Any] = diversity_loss_weight # ctc loss lowerCAmelCase_ : List[str] = ctc_loss_reduction lowerCAmelCase_ : Dict = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowerCAmelCase_ : Optional[int] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowerCAmelCase_ : Tuple = list(_lowerCamelCase ) lowerCAmelCase_ : Any = list(_lowerCamelCase ) lowerCAmelCase_ : Dict = list(_lowerCamelCase ) lowerCAmelCase_ : List[str] = xvector_output_dim @property def lowercase_ ( self ) -> Tuple: return functools.reduce(operator.mul , self.conv_stride , 1 )
262
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : Dict = logging.get_logger(__name__) _lowerCamelCase : Optional[Any] = ['model.decoder.embed_positions.weights'] def lowercase_ ( _UpperCAmelCase ): """simple docstring""" if "emb" in name: A_ : Tuple = name.replace('''emb''' , '''model.decoder.embed_tokens''' ) if "transformer" in name: A_ : Optional[int] = name.replace('''transformer''' , '''model.decoder''' ) if "cross_attention" in name: A_ : Optional[Any] = name.replace('''cross_attention''' , '''encoder_attn''' ) if "linear1" in name: A_ : int = name.replace('''linear1''' , '''fc1''' ) if "linear2" in name: A_ : Optional[int] = name.replace('''linear2''' , '''fc2''' ) if "norm1" in name: A_ : Any = name.replace('''norm1''' , '''self_attn_layer_norm''' ) if "norm_cross" in name: A_ : Any = name.replace('''norm_cross''' , '''encoder_attn_layer_norm''' ) if "norm2" in name: A_ : Dict = name.replace('''norm2''' , '''final_layer_norm''' ) if "out_norm" in name: A_ : Tuple = name.replace('''out_norm''' , '''model.decoder.layer_norm''' ) if "linears" in name: A_ : Union[str, Any] = name.replace('''linears''' , '''lm_heads''' ) if "condition_provider.conditioners.description.output_proj" in name: A_ : Tuple = name.replace('''condition_provider.conditioners.description.output_proj''' , '''enc_to_dec_proj''' ) return name def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = list(state_dict.keys() ) A_ : List[Any] = {} for key in keys: A_ : List[str] = state_dict.pop(_UpperCAmelCase ) A_ : Tuple = rename_keys(_UpperCAmelCase ) if "in_proj_weight" in key: # split fused qkv proj A_ : Any = val[:hidden_size, :] A_ : Optional[int] = val[hidden_size : 2 * hidden_size, :] A_ : Union[str, Any] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: A_ : List[str] = val else: A_ : int = val return state_dict, enc_dec_proj_state_dict def lowercase_ ( _UpperCAmelCase ): """simple docstring""" if checkpoint == "small": # default config values A_ : Optional[Any] = 1024 A_ : Tuple = 24 A_ : int = 16 elif checkpoint == "medium": A_ : Any = 1536 A_ : Union[str, Any] = 48 A_ : List[Any] = 24 elif checkpoint == "large": A_ : Optional[int] = 2048 A_ : Optional[int] = 48 A_ : Tuple = 32 else: raise ValueError(f"""Checkpoint should be one of `['small', 'medium', 'large']`, got {checkpoint}.""" ) A_ : Tuple = MusicgenDecoderConfig( hidden_size=_UpperCAmelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=_UpperCAmelCase , num_attention_heads=_UpperCAmelCase , ) return config @torch.no_grad() def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase="cpu" ): """simple docstring""" A_ : Any = MusicGen.get_pretrained(_UpperCAmelCase , device=_UpperCAmelCase ) A_ : str = decoder_config_from_checkpoint(_UpperCAmelCase ) A_ : Optional[int] = fairseq_model.lm.state_dict() A_ , A_ : str = rename_state_dict( _UpperCAmelCase , hidden_size=decoder_config.hidden_size ) A_ : List[str] = TaEncoderModel.from_pretrained('''t5-base''' ) A_ : Tuple = EncodecModel.from_pretrained('''facebook/encodec_32khz''' ) A_ : Union[str, Any] = MusicgenForCausalLM(_UpperCAmelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection A_ , A_ : Tuple = decoder.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) for key in missing_keys.copy(): if key.startswith(('''text_encoder''', '''audio_encoder''') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(_UpperCAmelCase ) if len(_UpperCAmelCase ) > 0: raise ValueError(f"""Missing key(s) in state_dict: {missing_keys}""" ) if len(_UpperCAmelCase ) > 0: raise ValueError(f"""Unexpected key(s) in state_dict: {unexpected_keys}""" ) # init the composite model A_ : Tuple = MusicgenForConditionalGeneration(text_encoder=_UpperCAmelCase , audio_encoder=_UpperCAmelCase , decoder=_UpperCAmelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(_UpperCAmelCase ) # check we can do a forward pass A_ : List[str] = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) A_ : Union[str, Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): A_ : Tuple = model(input_ids=_UpperCAmelCase , decoder_input_ids=_UpperCAmelCase ).logits if logits.shape != (8, 1, 2048): raise ValueError('''Incorrect shape for logits''' ) # now construct the processor A_ : str = AutoTokenizer.from_pretrained('''t5-base''' ) A_ : int = AutoFeatureExtractor.from_pretrained('''facebook/encodec_32khz''' , padding_side='''left''' ) A_ : Optional[int] = MusicgenProcessor(feature_extractor=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) # set the appropriate bos/pad token ids A_ : Tuple = 2048 A_ : Union[str, Any] = 2048 # set other default generation config params A_ : Union[str, Any] = int(30 * audio_encoder.config.frame_rate ) A_ : List[str] = True A_ : List[str] = 3.0 if pytorch_dump_folder is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) logger.info(f"""Saving model {checkpoint} to {pytorch_dump_folder}""" ) model.save_pretrained(_UpperCAmelCase ) processor.save_pretrained(_UpperCAmelCase ) if repo_id: logger.info(f"""Pushing model {checkpoint} to {repo_id}""" ) model.push_to_hub(_UpperCAmelCase ) processor.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": _lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) _lowerCamelCase : Optional[Any] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
167
0
'''simple docstring''' lowerCAmelCase : int = 2_56 # Modulus to hash a string lowerCAmelCase : Tuple = 1_00_00_03 def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : int = len(_A ) _lowerCAmelCase : Union[str, Any] = len(_A ) if p_len > t_len: return False _lowerCAmelCase : int = 0 _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : int = 1 # Calculating the hash of pattern and substring of text for i in range(_A ): _lowerCAmelCase : int = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus _lowerCAmelCase : int = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue _lowerCAmelCase : List[Any] = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash _lowerCAmelCase : Dict = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase (): """simple docstring""" _lowerCAmelCase : str = 'abc1abc12' _lowerCAmelCase : Union[str, Any] = 'alskfjaldsabc1abc1abc12k23adsfabcabc' _lowerCAmelCase : Union[str, Any] = 'alskfjaldsk23adsfabcabc' assert rabin_karp(_A , _A ) and not rabin_karp(_A , _A ) # Test 2) _lowerCAmelCase : Union[str, Any] = 'ABABX' _lowerCAmelCase : Any = 'ABABZABABYABABX' assert rabin_karp(_A , _A ) # Test 3) _lowerCAmelCase : List[Any] = 'AAAB' _lowerCAmelCase : int = 'ABAAAAAB' assert rabin_karp(_A , _A ) # Test 4) _lowerCAmelCase : Any = 'abcdabcy' _lowerCAmelCase : str = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(_A , _A ) # Test 5) _lowerCAmelCase : Any = 'Lü' _lowerCAmelCase : List[str] = 'Lüsai' assert rabin_karp(_A , _A ) _lowerCAmelCase : int = 'Lue' assert not rabin_karp(_A , _A ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
371
'''simple docstring''' import argparse import importlib from pathlib import Path # Test all the extensions added in the setup lowerCAmelCase : List[str] = [ """kernels/rwkv/wkv_cuda.cu""", """kernels/rwkv/wkv_op.cpp""", """kernels/deformable_detr/ms_deform_attn.h""", """kernels/deformable_detr/cuda/ms_deform_im2col_cuda.cuh""", """models/graphormer/algos_graphormer.pyx""", ] def lowercase (_A ): """simple docstring""" for file in FILES_TO_FIND: if not (transformers_path / file).exists(): return False return True if __name__ == "__main__": lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument("""--check_lib""", action="""store_true""", help="""Whether to check the build or the actual package.""") lowerCAmelCase : Dict = parser.parse_args() if args.check_lib: lowerCAmelCase : Union[str, Any] = importlib.import_module("""transformers""") lowerCAmelCase : int = Path(transformers_module.__file__).parent else: lowerCAmelCase : int = Path.cwd() / """build/lib/transformers""" if not test_custom_files_are_present(transformers_path): raise ValueError("""The built release does not contain the custom files. Fix this before going further!""")
25
0
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class a__ ( A__ , unittest.TestCase ): A = KandinskyVaaPriorPipeline A = ['prompt'] A = ['prompt', 'negative_prompt'] A = [ 'num_images_per_prompt', 'generator', 'num_inference_steps', 'latents', 'negative_prompt', 'guidance_scale', 'output_type', 'return_dict', ] A = False @property def __UpperCamelCase ( self : int ): """simple docstring""" return 32 @property def __UpperCamelCase ( self : int ): """simple docstring""" return 32 @property def __UpperCamelCase ( self : List[Any] ): """simple docstring""" return self.time_input_dim @property def __UpperCamelCase ( self : Dict ): """simple docstring""" return self.time_input_dim * 4 @property def __UpperCamelCase ( self : List[Any] ): """simple docstring""" return 100 @property def __UpperCamelCase ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) return tokenizer @property def __UpperCamelCase ( self : List[str] ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : str = CLIPTextConfig( bos_token_id=0,eos_token_id=2,hidden_size=self.text_embedder_hidden_size,projection_dim=self.text_embedder_hidden_size,intermediate_size=37,layer_norm_eps=1E-05,num_attention_heads=4,num_hidden_layers=5,pad_token_id=1,vocab_size=1000,) return CLIPTextModelWithProjection(_A ) @property def __UpperCamelCase ( self : Optional[Any] ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : Tuple = { "num_attention_heads": 2, "attention_head_dim": 12, "embedding_dim": self.text_embedder_hidden_size, "num_layers": 1, } SCREAMING_SNAKE_CASE_ : List[Any] = PriorTransformer(**_A ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 SCREAMING_SNAKE_CASE_ : Dict = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def __UpperCamelCase ( self : Any ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : Tuple = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size,image_size=224,projection_dim=self.text_embedder_hidden_size,intermediate_size=37,num_attention_heads=4,num_channels=3,num_hidden_layers=5,patch_size=14,) SCREAMING_SNAKE_CASE_ : Optional[Any] = CLIPVisionModelWithProjection(_A ) return model @property def __UpperCamelCase ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = CLIPImageProcessor( crop_size=224,do_center_crop=_A,do_normalize=_A,do_resize=_A,image_mean=[0.48145466, 0.4578275, 0.40821073],image_std=[0.26862954, 0.26130258, 0.27577711],resample=3,size=224,) return image_processor def __UpperCamelCase ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.dummy_prior SCREAMING_SNAKE_CASE_ : List[str] = self.dummy_image_encoder SCREAMING_SNAKE_CASE_ : Any = self.dummy_text_encoder SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.dummy_tokenizer SCREAMING_SNAKE_CASE_ : Optional[int] = self.dummy_image_processor SCREAMING_SNAKE_CASE_ : Any = UnCLIPScheduler( variance_type="fixed_small_log",prediction_type="sample",num_train_timesteps=1000,clip_sample=_A,clip_sample_range=10.0,) SCREAMING_SNAKE_CASE_ : str = { "prior": prior, "image_encoder": image_encoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "scheduler": scheduler, "image_processor": image_processor, } return components def __UpperCamelCase ( self : Tuple,_A : int,_A : Dict=0 ): """simple docstring""" if str(_A ).startswith("mps" ): SCREAMING_SNAKE_CASE_ : Tuple = torch.manual_seed(_A ) else: SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.Generator(device=_A ).manual_seed(_A ) SCREAMING_SNAKE_CASE_ : str = { "prompt": "horse", "generator": generator, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = "cpu" SCREAMING_SNAKE_CASE_ : List[Any] = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.pipeline_class(**_A ) SCREAMING_SNAKE_CASE_ : List[str] = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) SCREAMING_SNAKE_CASE_ : List[str] = pipe(**self.get_dummy_inputs(_A ) ) SCREAMING_SNAKE_CASE_ : int = output.image_embeds SCREAMING_SNAKE_CASE_ : Any = pipe( **self.get_dummy_inputs(_A ),return_dict=_A,)[0] SCREAMING_SNAKE_CASE_ : Any = image[0, -10:] SCREAMING_SNAKE_CASE_ : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) SCREAMING_SNAKE_CASE_ : Optional[int] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def __UpperCamelCase ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = torch_device == "cpu" SCREAMING_SNAKE_CASE_ : str = True SCREAMING_SNAKE_CASE_ : List[str] = False self._test_inference_batch_single_identical( test_max_difference=_A,relax_max_difference=_A,test_mean_pixel_difference=_A,) @skip_mps def __UpperCamelCase ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = torch_device == "cpu" SCREAMING_SNAKE_CASE_ : Dict = False self._test_attention_slicing_forward_pass( test_max_difference=_A,test_mean_pixel_difference=_A,)
18
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : List[str] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any=13 , __lowerCamelCase : Dict=3 , __lowerCamelCase : int=224 , __lowerCamelCase : Any=30 , __lowerCamelCase : Tuple=400 , __lowerCamelCase : int=True , __lowerCamelCase : List[str]=None , __lowerCamelCase : Any=True , __lowerCamelCase : Dict=[0.5, 0.5, 0.5] , __lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , ): UpperCamelCase :List[Any] = size if size is not None else {"""height""": 18, """width""": 18} UpperCamelCase :str = parent UpperCamelCase :Optional[int] = batch_size UpperCamelCase :Dict = num_channels UpperCamelCase :str = image_size UpperCamelCase :Dict = min_resolution UpperCamelCase :str = max_resolution UpperCamelCase :Union[str, Any] = do_resize UpperCamelCase :Optional[Any] = size UpperCamelCase :Any = do_normalize UpperCamelCase :Optional[Any] = image_mean UpperCamelCase :Tuple = image_std def _A ( self : int ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case__ : List[Any] = ViTImageProcessor if is_vision_available() else None def _A ( self : str ): UpperCamelCase :Tuple = EfficientFormerImageProcessorTester(self ) @property def _A ( self : List[str] ): return self.image_proc_tester.prepare_image_processor_dict() def _A ( self : int ): UpperCamelCase :List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) def _A ( self : Optional[int] ): pass def _A ( self : str ): # Initialize image_processor UpperCamelCase :Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase :Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , Image.Image ) # Test not batched input UpperCamelCase :List[str] = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched UpperCamelCase :List[Any] = image_processor(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def _A ( self : Union[str, Any] ): # Initialize image_processor UpperCamelCase :Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase :List[Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=__lowerCamelCase , numpify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , np.ndarray ) # Test not batched input UpperCamelCase :Dict = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched UpperCamelCase :Tuple = image_processor(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def _A ( self : List[Any] ): # Initialize image_processor UpperCamelCase :List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase :Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=__lowerCamelCase , torchify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , torch.Tensor ) # Test not batched input UpperCamelCase :List[Any] = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched UpperCamelCase :str = image_processor(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
38
0
'''simple docstring''' def lowercase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : int ): """simple docstring""" if not isinstance(lowercase__ , lowercase__ ): raise ValueError("""iterations must be defined as integers""" ) if not isinstance(lowercase__ , lowercase__ ) or not number >= 1: raise ValueError( """starting number must be and integer and be more than 0""" ) if not iterations >= 1: raise ValueError("""Iterations must be done more than 0 times to play FizzBuzz""" ) __UpperCAmelCase : str = """""" while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(lowercase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
355
'''simple docstring''' class _A : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase : int = data __UpperCAmelCase : int = previous __UpperCAmelCase : Union[str, Any] = next_node def __str__( self ) -> str: '''simple docstring''' return f'{self.data}' def __A ( self ) -> int: '''simple docstring''' return self.data def __A ( self ) -> List[str]: '''simple docstring''' return self.next def __A ( self ) -> str: '''simple docstring''' return self.previous class _A : def __init__( self , __UpperCAmelCase ) -> str: '''simple docstring''' __UpperCAmelCase : int = head def __iter__( self ) -> str: '''simple docstring''' return self def __A ( self ) -> str: '''simple docstring''' if not self.current: raise StopIteration else: __UpperCAmelCase : List[str] = self.current.get_data() __UpperCAmelCase : int = self.current.get_next() return value class _A : def __init__( self ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : Optional[Any] = None # First node in list __UpperCAmelCase : List[str] = None # Last node in list def __str__( self ) -> int: '''simple docstring''' __UpperCAmelCase : Tuple = self.head __UpperCAmelCase : Optional[int] = [] while current is not None: nodes.append(current.get_data() ) __UpperCAmelCase : Any = current.get_next() return " ".join(str(__UpperCAmelCase ) for node in nodes ) def __contains__( self , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : List[Any] = self.head while current: if current.get_data() == value: return True __UpperCAmelCase : Optional[Any] = current.get_next() return False def __iter__( self ) -> str: '''simple docstring''' return LinkedListIterator(self.head ) def __A ( self ) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def __A ( self ) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def __A ( self , __UpperCAmelCase ) -> None: '''simple docstring''' if self.head is None: __UpperCAmelCase : str = node __UpperCAmelCase : List[str] = node else: self.insert_before_node(self.head , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> None: '''simple docstring''' if self.head is None: self.set_head(__UpperCAmelCase ) else: self.insert_after_node(self.tail , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : Optional[int] = Node(__UpperCAmelCase ) if self.head is None: self.set_head(__UpperCAmelCase ) else: self.set_tail(__UpperCAmelCase ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : Tuple = node __UpperCAmelCase : List[Any] = node.previous if node.get_previous() is None: __UpperCAmelCase : str = node_to_insert else: __UpperCAmelCase : Optional[Any] = node_to_insert __UpperCAmelCase : List[Any] = node_to_insert def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : List[str] = node __UpperCAmelCase : Union[str, Any] = node.next if node.get_next() is None: __UpperCAmelCase : Dict = node_to_insert else: __UpperCAmelCase : Any = node_to_insert __UpperCAmelCase : List[str] = node_to_insert def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : Optional[Any] = 1 __UpperCAmelCase : Optional[Any] = Node(__UpperCAmelCase ) __UpperCAmelCase : Optional[Any] = self.head while node: if current_position == position: self.insert_before_node(__UpperCAmelCase , __UpperCAmelCase ) return current_position += 1 __UpperCAmelCase : int = node.next self.insert_after_node(self.tail , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> Node: '''simple docstring''' __UpperCAmelCase : Dict = self.head while node: if node.get_data() == item: return node __UpperCAmelCase : List[str] = node.get_next() raise Exception("""Node not found""" ) def __A ( self , __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' if (node := self.get_node(__UpperCAmelCase )) is not None: if node == self.head: __UpperCAmelCase : Optional[int] = self.head.get_next() if node == self.tail: __UpperCAmelCase : Union[str, Any] = self.tail.get_previous() self.remove_node_pointers(__UpperCAmelCase ) @staticmethod def __A ( __UpperCAmelCase ) -> None: '''simple docstring''' if node.get_next(): __UpperCAmelCase : Optional[Any] = node.previous if node.get_previous(): __UpperCAmelCase : int = node.next __UpperCAmelCase : Tuple = None __UpperCAmelCase : Union[str, Any] = None def __A ( self ) -> List[Any]: '''simple docstring''' return self.head is None def lowercase_ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
16
0
'''simple docstring''' import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = image.size __UpperCamelCase , __UpperCamelCase = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 __UpperCamelCase = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) __UpperCamelCase = np.array(__lowercase ).astype(np.floataa ) / 2_5_5.0 __UpperCamelCase = image[None].transpose(0 , 3 , 1 , 2 ) __UpperCamelCase = torch.from_numpy(__lowercase ) return 2.0 * image - 1.0 class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : str , __A : VQModel , __A : UNetaDModel , __A : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): super().__init__() self.register_modules(vqvae=__A , unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self : Any , __A : Union[torch.Tensor, PIL.Image.Image] = None , __A : Optional[int] = 1 , __A : Optional[int] = 1_0_0 , __A : Optional[float] = 0.0 , __A : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A : Optional[str] = "pil" , __A : bool = True , ): if isinstance(__A , PIL.Image.Image ): __UpperCamelCase = 1 elif isinstance(__A , torch.Tensor ): __UpperCamelCase = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(__A )}''' ) if isinstance(__A , PIL.Image.Image ): __UpperCamelCase = preprocess(__A ) __UpperCamelCase , __UpperCamelCase = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __UpperCamelCase = (batch_size, self.unet.config.in_channels // 2, height, width) __UpperCamelCase = next(self.unet.parameters() ).dtype __UpperCamelCase = randn_tensor(__A , generator=__A , device=self.device , dtype=__A ) __UpperCamelCase = image.to(device=self.device , dtype=__A ) # set timesteps and move to the correct device self.scheduler.set_timesteps(__A , device=self.device ) __UpperCamelCase = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __UpperCamelCase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __UpperCamelCase = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __UpperCamelCase = {} if accepts_eta: __UpperCamelCase = eta for t in self.progress_bar(__A ): # concat latents and low resolution image in the channel dimension. __UpperCamelCase = torch.cat([latents, image] , dim=1 ) __UpperCamelCase = self.scheduler.scale_model_input(__A , __A ) # predict the noise residual __UpperCamelCase = self.unet(__A , __A ).sample # compute the previous noisy sample x_t -> x_t-1 __UpperCamelCase = self.scheduler.step(__A , __A , __A , **__A ).prev_sample # decode the image latents with the VQVAE __UpperCamelCase = self.vqvae.decode(__A ).sample __UpperCamelCase = torch.clamp(__A , -1.0 , 1.0 ) __UpperCamelCase = image / 2 + 0.5 __UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __UpperCamelCase = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
53
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any]=False ) -> Tuple: """simple docstring""" try: __UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCamelCase = default else: # KEY is set, convert it to True or False. try: __UpperCamelCase = strtobool(__lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value a__ : str =parse_flag_from_env('''RUN_SLOW''', default=False) a__ : Union[str, Any] =parse_flag_from_env('''RUN_REMOTE''', default=False) a__ : List[str] =parse_flag_from_env('''RUN_LOCAL''', default=True) a__ : Optional[int] =parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression a__ : Any =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') a__ : Optional[int] =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') a__ : List[str] =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio a__ : Any =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam a__ : Tuple =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility a__ : Union[str, Any] =pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows a__ : int =pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase__ ( __lowercase : Optional[Any] ) -> Optional[int]: """simple docstring""" try: import faiss # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires faiss' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" try: import regex # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires regex' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Tuple ) -> List[Any]: """simple docstring""" try: import elasticsearch # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires elasticsearch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" try: import sqlalchemy # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires sqlalchemy' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[str] ) -> List[str]: """simple docstring""" if not config.TORCH_AVAILABLE: __UpperCamelCase = unittest.skip('test requires PyTorch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[Any] ) -> List[str]: """simple docstring""" if not config.TF_AVAILABLE: __UpperCamelCase = unittest.skip('test requires TensorFlow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" if not config.JAX_AVAILABLE: __UpperCamelCase = unittest.skip('test requires JAX' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> Optional[Any]: """simple docstring""" if not config.PIL_AVAILABLE: __UpperCamelCase = unittest.skip('test requires Pillow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('test requires transformers' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : int ) -> int: """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('test requires tiktoken' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> Any: """simple docstring""" def _require_spacy_model(__lowercase : Any ): try: import spacy # noqa F401 spacy.load(__lowercase ) except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) except OSError: return unittest.skip('test requires spacy model \'{}\''.format(__lowercase ) )(__lowercase ) else: return test_case return _require_spacy_model def lowercase__ ( __lowercase : Union[str, Any] ) -> str: """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('test requires pyspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('test requires joblibspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: __UpperCamelCase = unittest.skip('test is slow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_local_tests or _run_local_tests == 0: __UpperCamelCase = unittest.skip('test is local' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCamelCase = unittest.skip('test is packaged' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: __UpperCamelCase = unittest.skip('test requires remote' )(__lowercase ) return test_case def lowercase__ ( *__lowercase : Optional[Any] ) -> Tuple: """simple docstring""" def decorate(cls : int ): for name, fn in cls.__dict__.items(): if callable(__lowercase ) and name.startswith('test' ): for decorator in decorators: __UpperCamelCase = decorator(__lowercase ) setattr(cls , __lowercase , __lowercase ) return cls return decorate class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =0 SCREAMING_SNAKE_CASE_ : List[Any] =1 SCREAMING_SNAKE_CASE_ : Union[str, Any] =2 @contextmanager def lowercase__ ( __lowercase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , __lowercase : Dict=1e-16 ) -> List[Any]: """simple docstring""" __UpperCamelCase = requests.Session().request def timeout_request(__lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[Any] , **__lowercase : List[str] ): # Change the url to an invalid url so that the connection hangs __UpperCamelCase = 'https://10.255.255.1' if kwargs.get('timeout' ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCamelCase = timeout try: return online_request(__lowercase , __lowercase , **__lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCamelCase = url __UpperCamelCase = e.args[0] __UpperCamelCase = (max_retry_error.args[0].replace('10.255.255.1' , F'''OfflineMock[{url}]''' ),) __UpperCamelCase = (max_retry_error,) raise def raise_connection_error(__lowercase : int , __lowercase : List[str] , **__lowercase : Union[str, Any] ): raise requests.ConnectionError('Offline mode is enabled.' , request=__lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('requests.Session.send' , __lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('requests.Session.request' , __lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('datasets.config.HF_DATASETS_OFFLINE' , __lowercase ): yield else: raise ValueError('Please use a value from the OfflineSimulationMode enum.' ) @contextmanager def lowercase__ ( *__lowercase : Any , **__lowercase : Dict ) -> Dict: """simple docstring""" __UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__lowercase , **__lowercase ) as tmp_dir: try: os.chdir(__lowercase ) yield finally: os.chdir(__lowercase ) @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase__ ( __lowercase : List[str] , __lowercase : int ) -> Union[str, Any]: """simple docstring""" return deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() == deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(__lowercase : List[Any] , *__lowercase : Tuple , **__lowercase : Union[str, Any] ): try: return func(*__lowercase , **__lowercase ) except HTTPError as err: if str(__lowercase ).startswith('500' ) or str(__lowercase ).startswith('502' ): pytest.xfail(str(__lowercase ) ) raise err return decorator.decorator(_wrapper , __lowercase ) class snake_case : """simple docstring""" def __init__( self : int , __A : Any , __A : str , __A : List[Any] ): __UpperCamelCase = returncode __UpperCamelCase = stdout __UpperCamelCase = stderr async def lowercase__ ( __lowercase : Any , __lowercase : Optional[int] ) -> str: """simple docstring""" while True: __UpperCamelCase = await stream.readline() if line: callback(__lowercase ) else: break async def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any]=None , __lowercase : Any=None , __lowercase : Optional[Any]=None , __lowercase : int=False , __lowercase : List[Any]=False ) -> _RunOutput: """simple docstring""" if echo: print('\nRunning: ' , ' '.join(__lowercase ) ) __UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=__lowercase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=__lowercase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCamelCase = [] __UpperCamelCase = [] def tee(__lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : Tuple="" ): __UpperCamelCase = line.decode('utf-8' ).rstrip() sink.append(__lowercase ) if not quiet: print(__lowercase , __lowercase , file=__lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda __lowercase : tee(__lowercase , __lowercase , sys.stdout , label='stdout:' ) ), _read_stream(p.stderr , lambda __lowercase : tee(__lowercase , __lowercase , sys.stderr , label='stderr:' ) ), ] , timeout=__lowercase , ) return _RunOutput(await p.wait() , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict , __lowercase : Any=None , __lowercase : int=None , __lowercase : int=180 , __lowercase : int=False , __lowercase : str=True ) -> _RunOutput: """simple docstring""" __UpperCamelCase = asyncio.get_event_loop() __UpperCamelCase = loop.run_until_complete( _stream_subprocess(__lowercase , env=__lowercase , stdin=__lowercase , timeout=__lowercase , quiet=__lowercase , echo=__lowercase ) ) __UpperCamelCase = ' '.join(__lowercase ) if result.returncode > 0: __UpperCamelCase = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def lowercase__ ( ) -> List[str]: """simple docstring""" __UpperCamelCase = os.environ.get('PYTEST_XDIST_WORKER' , 'gw0' ) __UpperCamelCase = re.sub(R'^gw' , '' , __lowercase , 0 , re.M ) return int(__lowercase ) def lowercase__ ( ) -> List[Any]: """simple docstring""" __UpperCamelCase = 29500 __UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
53
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE__ = {"""configuration_mra""": ["""MRA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MraConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ """MRA_PRETRAINED_MODEL_ARCHIVE_LIST""", """MraForMaskedLM""", """MraForMultipleChoice""", """MraForQuestionAnswering""", """MraForSequenceClassification""", """MraForTokenClassification""", """MraLayer""", """MraModel""", """MraPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
362
import sys def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Optional[Any] ): '''simple docstring''' lowercase_ = len(__lowerCamelCase ) lowercase_ = [[0 for x in range(__lowerCamelCase )] for x in range(__lowerCamelCase )] lowercase_ = [[0 for x in range(__lowerCamelCase )] for x in range(__lowerCamelCase )] for chain_length in range(2 , __lowerCamelCase ): for a in range(1 , n - chain_length + 1 ): lowercase_ = a + chain_length - 1 lowercase_ = sys.maxsize for c in range(__lowerCamelCase , __lowerCamelCase ): lowercase_ = ( matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b] ) if cost < matrix[a][b]: lowercase_ = cost lowercase_ = c return matrix, sol def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Optional[Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: Dict ): '''simple docstring''' if i == j: print("A" + str(__lowerCamelCase ) , end=" " ) else: print("(" , end=" " ) print_optiomal_solution(__lowerCamelCase , __lowerCamelCase , optimal_solution[i][j] ) print_optiomal_solution(__lowerCamelCase , optimal_solution[i][j] + 1 , __lowerCamelCase ) print(")" , end=" " ) def SCREAMING_SNAKE_CASE_ ( ): '''simple docstring''' lowercase_ = [30, 35, 15, 5, 10, 20, 25] lowercase_ = len(__lowerCamelCase ) # Size of matrix created from above array will be # 30*35 35*15 15*5 5*10 10*20 20*25 lowercase_ , lowercase_ = matrix_chain_order(__lowerCamelCase ) print("No. of Operation required: " + str(matrix[1][n - 1] ) ) print_optiomal_solution(__lowerCamelCase , 1 , n - 1 ) if __name__ == "__main__": main()
297
0
from ...configuration_utils import PretrainedConfig class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "bert-generation" def __init__( self , UpperCAmelCase=5_0358 , UpperCAmelCase=1024 , UpperCAmelCase=24 , UpperCAmelCase=16 , UpperCAmelCase=4096 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-12 , UpperCAmelCase=0 , UpperCAmelCase=2 , UpperCAmelCase=1 , UpperCAmelCase="absolute" , UpperCAmelCase=True , **UpperCAmelCase , ): """simple docstring""" super().__init__(pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = hidden_act _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = initializer_range _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = position_embedding_type _UpperCAmelCase = use_cache
39
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase__ : def __init__(self , UpperCAmelCase ) -> Any: _lowercase =str(id_ ) _lowercase =None _lowercase =None _lowercase =[] _lowercase ={} # {vertex:distance} def __lt__(self , UpperCAmelCase ) -> List[str]: return self.key < other.key def __repr__(self ) -> str: return self.id def __A (self , UpperCAmelCase ) -> Dict: self.neighbors.append(UpperCAmelCase ) def __A (self , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: _lowercase =weight def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case , __snake_case ) -> List[str]: """simple docstring""" graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __snake_case ) graph[b - 1].add_edge(graph[a - 1] , __snake_case ) def UpperCAmelCase_ ( __snake_case , __snake_case ) -> list: """simple docstring""" _lowercase =[] for u in graph: _lowercase =math.inf _lowercase =None _lowercase =0 _lowercase =graph[:] while q: _lowercase =min(__snake_case ) q.remove(__snake_case ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _lowercase =u _lowercase =u.edges[v.id] for i in range(1 , len(__snake_case ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCAmelCase_ ( __snake_case , __snake_case ) -> Iterator[tuple]: """simple docstring""" for u in graph: _lowercase =math.inf _lowercase =None _lowercase =0 _lowercase =list(__snake_case ) hq.heapify(__snake_case ) while h: _lowercase =hq.heappop(__snake_case ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _lowercase =u _lowercase =u.edges[v.id] hq.heapify(__snake_case ) for i in range(1 , len(__snake_case ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCAmelCase_ ( ) -> None: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
5
0
"""simple docstring""" from __future__ import annotations import collections import pprint from pathlib import Path def UpperCamelCase ( __lowercase : str ): '''simple docstring''' return "".join(sorted(__lowercase ) ) def UpperCamelCase ( __lowercase : str ): '''simple docstring''' return word_by_signature[signature(__lowercase )] _UpperCAmelCase = Path(__file__).parent.joinpath("""words.txt""").read_text(encoding="""utf-8""") _UpperCAmelCase = sorted({word.strip().lower() for word in data.splitlines()}) _UpperCAmelCase = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _UpperCAmelCase = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("""anagrams.txt""", """w""") as file: file.write("""all_anagrams = \n """) file.write(pprint.pformat(all_anagrams))
359
from math import isqrt def UpperCamelCase ( __lowercase : int ): '''simple docstring''' return all(number % divisor != 0 for divisor in range(2 ,isqrt(__lowercase ) + 1 ) ) def UpperCamelCase ( __lowercase : int = 10**6 ): '''simple docstring''' A_ : Optional[Any] = 0 A_ : List[str] = 1 A_ : Dict = 7 while prime_candidate < max_prime: primes_count += is_prime(__lowercase ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
192
0
import os from math import logaa def SCREAMING_SNAKE_CASE ( lowercase_ = "base_exp.txt" ) -> Any: """simple docstring""" A__ = 0 A__ = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(lowercase_ ) , lowercase_ ) ) ): A__ = list(map(lowercase_ , line.split(''',''' ) ) ) if x * logaa(lowercase_ ) > largest: A__ = x * logaa(lowercase_ ) A__ = i + 1 return result if __name__ == "__main__": print(solution())
14
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax _lowerCAmelCase = logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE_ ) class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' def __init__( self ,**__UpperCAmelCase ) -> Tuple: super().__init__(**__UpperCAmelCase ) requires_backends(self ,"""vision""" ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self ,__UpperCAmelCase ,**__UpperCAmelCase ) -> str: return super().__call__(__UpperCAmelCase ,**__UpperCAmelCase ) def UpperCAmelCase_ ( self ,**__UpperCAmelCase ) -> str: lowerCAmelCase__ : List[Any] = {} if "candidate_labels" in kwargs: lowerCAmelCase__ : int = kwargs["""candidate_labels"""] if "hypothesis_template" in kwargs: lowerCAmelCase__ : Optional[int] = kwargs["""hypothesis_template"""] return preprocess_params, {}, {} def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase=None ,__UpperCAmelCase="This is a photo of {}." ) -> int: lowerCAmelCase__ : str = load_image(__UpperCAmelCase ) lowerCAmelCase__ : Dict = self.image_processor(images=[image] ,return_tensors=self.framework ) lowerCAmelCase__ : List[Any] = candidate_labels lowerCAmelCase__ : List[str] = [hypothesis_template.format(__UpperCAmelCase ) for x in candidate_labels] lowerCAmelCase__ : Optional[Any] = self.tokenizer(__UpperCAmelCase ,return_tensors=self.framework ,padding=__UpperCAmelCase ) lowerCAmelCase__ : Tuple = [text_inputs] return inputs def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Union[str, Any]: lowerCAmelCase__ : Tuple = model_inputs.pop("""candidate_labels""" ) lowerCAmelCase__ : Union[str, Any] = model_inputs.pop("""text_inputs""" ) if isinstance(text_inputs[0] ,__UpperCAmelCase ): lowerCAmelCase__ : int = text_inputs[0] else: # Batching case. lowerCAmelCase__ : Dict = text_inputs[0][0] lowerCAmelCase__ : Any = self.model(**__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = { """candidate_labels""": candidate_labels, """logits""": outputs.logits_per_image, } return model_outputs def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Any: lowerCAmelCase__ : Union[str, Any] = model_outputs.pop("""candidate_labels""" ) lowerCAmelCase__ : List[str] = model_outputs["""logits"""][0] if self.framework == "pt": lowerCAmelCase__ : List[str] = logits.softmax(dim=-1 ).squeeze(-1 ) lowerCAmelCase__ : Optional[Any] = probs.tolist() if not isinstance(__UpperCAmelCase ,__UpperCAmelCase ): lowerCAmelCase__ : Dict = [scores] elif self.framework == "tf": lowerCAmelCase__ : Any = stable_softmax(__UpperCAmelCase ,axis=-1 ) lowerCAmelCase__ : List[Any] = probs.numpy().tolist() else: raise ValueError(F"""Unsupported framework: {self.framework}""" ) lowerCAmelCase__ : Tuple = [ {"""score""": score, """label""": candidate_label} for score, candidate_label in sorted(zip(__UpperCAmelCase ,__UpperCAmelCase ) ,key=lambda __UpperCAmelCase : -x[0] ) ] return result
37
0
import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class lowerCAmelCase ( unittest.TestCase ): def __init__( self :Dict , _lowercase :int , _lowercase :List[Any]=1_00 , _lowercase :Optional[int]=13 , _lowercase :List[Any]=30 , _lowercase :Union[str, Any]=2 , _lowercase :int=3 , _lowercase :List[str]=True , _lowercase :Any=True , _lowercase :Tuple=32 , _lowercase :Tuple=5 , _lowercase :List[Any]=4 , _lowercase :List[Any]=37 , _lowercase :int="gelu" , _lowercase :List[Any]=0.1 , _lowercase :str=0.1 , _lowercase :Tuple=10 , _lowercase :List[Any]=0.02 , _lowercase :List[str]=3 , ): '''simple docstring''' lowercase__ = parent lowercase__ = vocab_size lowercase__ = batch_size lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = is_training lowercase__ = use_labels lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = type_sequence_label_size lowercase__ = initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowercase__ = (image_size // patch_size) ** 2 lowercase__ = num_patches + 1 def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, pixel_values, labels def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Any , _lowercase :List[Any] ): '''simple docstring''' lowercase__ = FlaxBeitModel(config=_SCREAMING_SNAKE_CASE ) lowercase__ = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :int , _lowercase :int ): '''simple docstring''' lowercase__ = FlaxBeitForMaskedImageModeling(config=_SCREAMING_SNAKE_CASE ) lowercase__ = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :List[Any] , _lowercase :Dict , _lowercase :Any ): '''simple docstring''' lowercase__ = self.type_sequence_label_size lowercase__ = FlaxBeitForImageClassification(config=_SCREAMING_SNAKE_CASE ) lowercase__ = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowercase__ = 1 lowercase__ = FlaxBeitForImageClassification(_SCREAMING_SNAKE_CASE ) lowercase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase__ = model(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() ( ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ) = config_and_inputs lowercase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): __lowerCamelCase = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = FlaxBeitModelTester(self ) lowercase__ = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=37 ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_SCREAMING_SNAKE_CASE ) lowercase__ = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowercase__ = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowercase__ = model_class(_SCREAMING_SNAKE_CASE ) @jax.jit def model_jitted(_lowercase :Union[str, Any] , **_lowercase :Optional[int] ): return model(pixel_values=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) with self.subTest("JIT Enabled" ): lowercase__ = model_jitted(**_SCREAMING_SNAKE_CASE ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): lowercase__ = model_jitted(**_SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) ) for jitted_output, output in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: lowercase__ = model_class_name.from_pretrained("microsoft/beit-base-patch16-224" ) lowercase__ = model(np.ones((1, 3, 2_24, 2_24) ) ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def _A ( ): lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @require_flax class lowerCAmelCase ( unittest.TestCase ): @cached_property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return BeitImageProcessor.from_pretrained("microsoft/beit-base-patch16-224" ) if is_vision_available() else None @slow def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = FlaxBeitForMaskedImageModeling.from_pretrained("microsoft/beit-base-patch16-224-pt22k" ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="np" ).pixel_values # prepare bool_masked_pos lowercase__ = np.ones((1, 1_96) , dtype=_SCREAMING_SNAKE_CASE ) # forward pass lowercase__ = model(pixel_values=_SCREAMING_SNAKE_CASE , bool_masked_pos=_SCREAMING_SNAKE_CASE ) lowercase__ = outputs.logits # verify the logits lowercase__ = (1, 1_96, 81_92) self.assertEqual(logits.shape , _SCREAMING_SNAKE_CASE ) lowercase__ = np.array( [[-3.2437, 0.5072, -13.91_74], [-3.2456, 0.4948, -13.94_01], [-3.2033, 0.5121, -13.85_50]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-2 ) ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-base-patch16-224" ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="np" ) # forward pass lowercase__ = model(**_SCREAMING_SNAKE_CASE ) lowercase__ = outputs.logits # verify the logits lowercase__ = (1, 10_00) self.assertEqual(logits.shape , _SCREAMING_SNAKE_CASE ) lowercase__ = np.array([-1.2385, -1.0987, -1.0108] ) self.assertTrue(np.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) lowercase__ = 2_81 self.assertEqual(logits.argmax(-1 ).item() , _SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-large-patch16-224-pt22k-ft22k" ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="np" ) # forward pass lowercase__ = model(**_SCREAMING_SNAKE_CASE ) lowercase__ = outputs.logits # verify the logits lowercase__ = (1, 2_18_41) self.assertEqual(logits.shape , _SCREAMING_SNAKE_CASE ) lowercase__ = np.array([1.6881, -0.2787, 0.5901] ) self.assertTrue(np.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) lowercase__ = 23_96 self.assertEqual(logits.argmax(-1 ).item() , _SCREAMING_SNAKE_CASE )
355
from __future__ import annotations def _A ( __magic_name__ , __magic_name__ ): lowercase__ = [] create_all_state(1 , __magic_name__ , __magic_name__ , [] , __magic_name__ ) return result def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ): if level == 0: total_list.append(current_list[:] ) return for i in range(__magic_name__ , total_number - level + 2 ): current_list.append(__magic_name__ ) create_all_state(i + 1 , __magic_name__ , level - 1 , __magic_name__ , __magic_name__ ) current_list.pop() def _A ( __magic_name__ ): for i in total_list: print(*__magic_name__ ) if __name__ == "__main__": _snake_case = 4 _snake_case = 2 _snake_case = generate_all_combinations(n, k) print_all_state(total_list)
201
0
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCamelCase (_snake_case ): '''simple docstring''' def __init__( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = False , _UpperCamelCase = False , _UpperCamelCase = None , _UpperCamelCase = None , **_UpperCamelCase , ) -> Optional[int]: super().__init__( _UpperCamelCase , split=_UpperCamelCase , features=_UpperCamelCase , cache_dir=_UpperCamelCase , keep_in_memory=_UpperCamelCase , streaming=_UpperCamelCase , num_proc=_UpperCamelCase , **_UpperCamelCase , ) UpperCAmelCase_ : Any = field UpperCAmelCase_ : Tuple = path_or_paths if isinstance(_UpperCamelCase , _UpperCamelCase ) else {self.split: path_or_paths} UpperCAmelCase_ : int = Json( cache_dir=_UpperCamelCase , data_files=_UpperCamelCase , features=_UpperCamelCase , field=_UpperCamelCase , **_UpperCamelCase , ) def __UpperCAmelCase ( self ) -> List[Any]: # Build iterable dataset if self.streaming: UpperCAmelCase_ : Optional[Any] = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: UpperCAmelCase_ : Optional[int] = None UpperCAmelCase_ : List[Any] = None UpperCAmelCase_ : Any = None UpperCAmelCase_ : Any = None self.builder.download_and_prepare( download_config=_UpperCamelCase , download_mode=_UpperCamelCase , verification_mode=_UpperCamelCase , base_path=_UpperCamelCase , num_proc=self.num_proc , ) UpperCAmelCase_ : List[str] = self.builder.as_dataset( split=self.split , verification_mode=_UpperCamelCase , in_memory=self.keep_in_memory ) return dataset class lowerCamelCase : '''simple docstring''' def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , **_UpperCamelCase , ) -> Union[str, Any]: if num_proc is not None and num_proc <= 0: raise ValueError(f"num_proc {num_proc} must be an integer > 0." ) UpperCAmelCase_ : List[Any] = dataset UpperCAmelCase_ : Optional[int] = path_or_buf UpperCAmelCase_ : List[str] = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE UpperCAmelCase_ : str = num_proc UpperCAmelCase_ : Optional[int] = 'utf-8' UpperCAmelCase_ : Optional[int] = to_json_kwargs def __UpperCAmelCase ( self ) -> int: UpperCAmelCase_ : Dict = self.to_json_kwargs.pop('path_or_buf' , _UpperCamelCase ) UpperCAmelCase_ : Union[str, Any] = self.to_json_kwargs.pop('orient' , 'records' ) UpperCAmelCase_ : Any = self.to_json_kwargs.pop('lines' , True if orient == 'records' else False ) UpperCAmelCase_ : str = self.to_json_kwargs.pop('index' , False if orient in ['split', 'table'] else True ) UpperCAmelCase_ : List[str] = self.to_json_kwargs.pop('compression' , _UpperCamelCase ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(f"`datasets` currently does not support {compression} compression" ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , 'wb' , compression=_UpperCamelCase ) as buffer: UpperCAmelCase_ : Any = self._write(file_obj=_UpperCamelCase , orient=_UpperCamelCase , lines=_UpperCamelCase , index=_UpperCamelCase , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( f"The compression parameter is not supported when writing to a buffer, but compression={compression}" ' was passed. Please provide a local path instead.' ) UpperCAmelCase_ : Tuple = self._write( file_obj=self.path_or_buf , orient=_UpperCamelCase , lines=_UpperCamelCase , index=_UpperCamelCase , **self.to_json_kwargs ) return written def __UpperCAmelCase ( self , _UpperCamelCase ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[str] = args UpperCAmelCase_ : List[str] = query_table( table=self.dataset.data , key=slice(_UpperCamelCase , offset + self.batch_size ) , indices=self.dataset._indices , ) UpperCAmelCase_ : str = batch.to_pandas().to_json( path_or_buf=_UpperCamelCase , orient=_UpperCamelCase , lines=_UpperCamelCase , index=_UpperCamelCase , **_UpperCamelCase ) if not json_str.endswith('\n' ): json_str += "\n" return json_str.encode(self.encoding ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase , ) -> int: UpperCAmelCase_ : Any = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating json from Arrow format' , ): UpperCAmelCase_ : List[Any] = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(_UpperCamelCase ) else: UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , _UpperCamelCase , _UpperCamelCase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating json from Arrow format' , ): written += file_obj.write(_UpperCamelCase ) return written
29
import os # Precomputes a list of the 100 first triangular numbers __UpperCAmelCase = [int(0.5 * n * (n + 1)) for n in range(1, 101)] def lowercase__ ( ): '''simple docstring''' UpperCAmelCase_ : Any = os.path.dirname(os.path.realpath(__snake_case ) ) UpperCAmelCase_ : Optional[Any] = os.path.join(__snake_case , 'words.txt' ) UpperCAmelCase_ : Union[str, Any] = '' with open(__snake_case ) as f: UpperCAmelCase_ : List[Any] = f.readline() UpperCAmelCase_ : Optional[int] = [word.strip('"' ) for word in words.strip('\r\n' ).split(',' )] UpperCAmelCase_ : Optional[int] = [ word for word in [sum(ord(__snake_case ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(__snake_case ) if __name__ == "__main__": print(solution())
29
1
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo snake_case__ : int = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey\n and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin\n Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto\n Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and\n Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes\n and Jeffrey Dean},\n year={2016},\n eprint={1609.08144},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n' snake_case__ : Optional[int] = '\\nThe BLEU score has some undesirable properties when used for single\nsentences, as it was designed to be a corpus measure. We therefore\nuse a slightly different score for our RL experiments which we call\nthe \'GLEU score\'. For the GLEU score, we record all sub-sequences of\n1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then\ncompute a recall, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the target (ground truth) sequence,\nand a precision, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the generated output sequence. Then\nGLEU score is simply the minimum of recall and precision. This GLEU\nscore\'s range is always between 0 (no matches) and 1 (all match) and\nit is symmetrical when switching output and target. According to\nour experiments, GLEU score correlates quite well with the BLEU\nmetric on a corpus level but does not have its drawbacks for our per\nsentence reward objective.\n' snake_case__ : Optional[int] = '\\nComputes corpus-level Google BLEU (GLEU) score of translated segments against one or more references.\nInstead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching\ntokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values.\n\nArgs:\n predictions (list of str): list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references (list of list of str): list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n min_len (int): The minimum order of n-gram this function should extract. Defaults to 1.\n max_len (int): The maximum order of n-gram this function should extract. Defaults to 4.\n\nReturns:\n \'google_bleu\': google_bleu score\n\nExamples:\n Example 1:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.44\n\n Example 2:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.61\n\n Example 3:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2)\n >>> print(round(results["google_bleu"], 2))\n 0.53\n\n Example 4:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6)\n >>> print(round(results["google_bleu"], 2))\n 0.4\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ ( datasets.Metric ): def _lowerCAmelCase (self :Union[str, Any] )-> MetricInfo: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ), '''references''': datasets.Sequence( datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ) , id='''references''' ), } ) , ) def _lowerCAmelCase (self :List[str] , _UpperCamelCase :List[List[List[str]]] , _UpperCamelCase :List[List[str]] , _UpperCamelCase :int = 1 , _UpperCamelCase :int = 4 , )-> Dict[str, float]: return { "google_bleu": gleu_score.corpus_gleu( list_of_references=_UpperCamelCase , hypotheses=_UpperCamelCase , min_len=_UpperCamelCase , max_len=_UpperCamelCase ) }
371
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ : List[str] = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ : Optional[Any] = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys snake_case__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
250
0
from string import ascii_lowercase, ascii_uppercase def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: str ) -> str: '''simple docstring''' if not sentence: return "" A__ = dict(zip(__lowerCAmelCase , __lowerCAmelCase ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
68
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available SCREAMING_SNAKE_CASE__ : Optional[int] = logging.getLogger(__name__) @dataclass class lowerCAmelCase__ : a__ : str a__ : List[str] a__ : Optional[List[str]] @dataclass class lowerCAmelCase__ : a__ : List[int] a__ : List[int] a__ : Optional[List[int]] = None a__ : Optional[List[int]] = None class lowerCAmelCase__ ( __lowercase ): a__ : Optional[Any] = """train""" a__ : Optional[int] = """dev""" a__ : Dict = """test""" class lowerCAmelCase__ : @staticmethod def __A ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[Split, str] ) -> List[InputExample]: raise NotImplementedError @staticmethod def __A ( SCREAMING_SNAKE_CASE__ : str ) -> List[str]: raise NotImplementedError @staticmethod def __A ( SCREAMING_SNAKE_CASE__ : List[InputExample] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : List[str]="[CLS]" , SCREAMING_SNAKE_CASE__ : Tuple=1 , SCREAMING_SNAKE_CASE__ : str="[SEP]" , SCREAMING_SNAKE_CASE__ : List[Any]=False , SCREAMING_SNAKE_CASE__ : Union[str, Any]=False , SCREAMING_SNAKE_CASE__ : Tuple=0 , SCREAMING_SNAKE_CASE__ : int=0 , SCREAMING_SNAKE_CASE__ : str=-1_00 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0 , SCREAMING_SNAKE_CASE__ : List[Any]=True , ) -> List[InputFeatures]: __lowerCamelCase = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE__ )} __lowerCamelCase = [] for ex_index, example in enumerate(SCREAMING_SNAKE_CASE__ ): if ex_index % 1_00_00 == 0: logger.info('''Writing example %d of %d''' , SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) __lowerCamelCase = [] __lowerCamelCase = [] for word, label in zip(example.words , example.labels ): __lowerCamelCase = tokenizer.tokenize(SCREAMING_SNAKE_CASE__ ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(SCREAMING_SNAKE_CASE__ ) > 0: tokens.extend(SCREAMING_SNAKE_CASE__ ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(SCREAMING_SNAKE_CASE__ ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. __lowerCamelCase = tokenizer.num_special_tokens_to_add() if len(SCREAMING_SNAKE_CASE__ ) > max_seq_length - special_tokens_count: __lowerCamelCase = tokens[: (max_seq_length - special_tokens_count)] __lowerCamelCase = label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] __lowerCamelCase = [sequence_a_segment_id] * len(SCREAMING_SNAKE_CASE__ ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: __lowerCamelCase = [cls_token] + tokens __lowerCamelCase = [pad_token_label_id] + label_ids __lowerCamelCase = [cls_token_segment_id] + segment_ids __lowerCamelCase = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. __lowerCamelCase = [1 if mask_padding_with_zero else 0] * len(SCREAMING_SNAKE_CASE__ ) # Zero-pad up to the sequence length. __lowerCamelCase = max_seq_length - len(SCREAMING_SNAKE_CASE__ ) if pad_on_left: __lowerCamelCase = ([pad_token] * padding_length) + input_ids __lowerCamelCase = ([0 if mask_padding_with_zero else 1] * padding_length) + input_mask __lowerCamelCase = ([pad_token_segment_id] * padding_length) + segment_ids __lowerCamelCase = ([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(SCREAMING_SNAKE_CASE__ ) == max_seq_length assert len(SCREAMING_SNAKE_CASE__ ) == max_seq_length assert len(SCREAMING_SNAKE_CASE__ ) == max_seq_length assert len(SCREAMING_SNAKE_CASE__ ) == max_seq_length if ex_index < 5: logger.info('''*** Example ***''' ) logger.info('''guid: %s''' , example.guid ) logger.info('''tokens: %s''' , ''' '''.join([str(SCREAMING_SNAKE_CASE__ ) for x in tokens] ) ) logger.info('''input_ids: %s''' , ''' '''.join([str(SCREAMING_SNAKE_CASE__ ) for x in input_ids] ) ) logger.info('''input_mask: %s''' , ''' '''.join([str(SCREAMING_SNAKE_CASE__ ) for x in input_mask] ) ) logger.info('''segment_ids: %s''' , ''' '''.join([str(SCREAMING_SNAKE_CASE__ ) for x in segment_ids] ) ) logger.info('''label_ids: %s''' , ''' '''.join([str(SCREAMING_SNAKE_CASE__ ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: __lowerCamelCase = None features.append( InputFeatures( input_ids=SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , label_ids=SCREAMING_SNAKE_CASE__ ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class lowerCAmelCase__ ( __lowercase ): a__ : List[InputFeatures] a__ : int = nn.CrossEntropyLoss().ignore_index def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : TokenClassificationTask , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[int] = None , SCREAMING_SNAKE_CASE__ : Union[str, Any]=False , SCREAMING_SNAKE_CASE__ : Split = Split.train , ) -> Union[str, Any]: # Load data features from cache or dataset file __lowerCamelCase = os.path.join( SCREAMING_SNAKE_CASE__ , '''cached_{}_{}_{}'''.format(mode.value , tokenizer.__class__.__name__ , str(SCREAMING_SNAKE_CASE__ ) ) , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowerCamelCase = cached_features_file + '''.lock''' with FileLock(SCREAMING_SNAKE_CASE__ ): if os.path.exists(SCREAMING_SNAKE_CASE__ ) and not overwrite_cache: logger.info(f'''Loading features from cached file {cached_features_file}''' ) __lowerCamelCase = torch.load(SCREAMING_SNAKE_CASE__ ) else: logger.info(f'''Creating features from dataset file at {data_dir}''' ) __lowerCamelCase = token_classification_task.read_examples_from_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # TODO clean up all this to leverage built-in features of tokenizers __lowerCamelCase = token_classification_task.convert_examples_to_features( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , cls_token_at_end=bool(model_type in ['''xlnet'''] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ['''xlnet'''] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=SCREAMING_SNAKE_CASE__ , pad_on_left=bool(tokenizer.padding_side == '''left''' ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info(f'''Saving features into cached file {cached_features_file}''' ) torch.save(self.features , SCREAMING_SNAKE_CASE__ ) def __len__( self : Dict ) -> str: return len(self.features ) def __getitem__( self : Any , SCREAMING_SNAKE_CASE__ : Dict ) -> InputFeatures: return self.features[i] if is_tf_available(): import tensorflow as tf class lowerCAmelCase__ : a__ : List[InputFeatures] a__ : int = -100 def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : TokenClassificationTask , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[int] = None , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Split = Split.train , ) -> List[Any]: __lowerCamelCase = token_classification_task.read_examples_from_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # TODO clean up all this to leverage built-in features of tokenizers __lowerCamelCase = token_classification_task.convert_examples_to_features( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , cls_token_at_end=bool(model_type in ['''xlnet'''] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ['''xlnet'''] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=SCREAMING_SNAKE_CASE__ , pad_on_left=bool(tokenizer.padding_side == '''left''' ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: __lowerCamelCase = tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE__ , ({'''input_ids''': tf.intaa, '''attention_mask''': tf.intaa}, tf.intaa) , ( {'''input_ids''': tf.TensorShape([None] ), '''attention_mask''': tf.TensorShape([None] )}, tf.TensorShape([None] ), ) , ) else: __lowerCamelCase = tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE__ , ({'''input_ids''': tf.intaa, '''attention_mask''': tf.intaa, '''token_type_ids''': tf.intaa}, tf.intaa) , ( { '''input_ids''': tf.TensorShape([None] ), '''attention_mask''': tf.TensorShape([None] ), '''token_type_ids''': tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) , ) def __A ( self : Union[str, Any] ) -> Union[str, Any]: __lowerCamelCase = self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self : List[Any] ) -> Any: return len(self.features ) def __getitem__( self : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> InputFeatures: return self.features[i]
270
0
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowerCamelCase__ ) , "Tatoeba directory does not exist." ) class lowercase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase__ : List[Any] = tempfile.mkdtemp() return TatoebaConverter(save_dir=__snake_case ) @slow def UpperCamelCase__ ( self ) -> List[Any]: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCamelCase__ ( self ) -> int: """simple docstring""" UpperCamelCase__ : Optional[int] = self.resolver.write_model_card('''opus-mt-he-en''', dry_run=__snake_case ) assert mmeta["long_pair"] == "heb-eng"
358
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available UpperCAmelCase_ = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
247
0
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor SCREAMING_SNAKE_CASE : str = random.Random() def lowercase ( _snake_case : Union[str, Any] , _snake_case : Optional[int]=1.0 , _snake_case : str=None , _snake_case : Tuple=None ) ->Any: """simple docstring""" if rng is None: __snake_case : List[str] = global_rng __snake_case : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__(self , a_ , a_=7 , a_=4_00 , a_=20_00 , a_=24 , a_=24 , a_=0.0 , a_=1_60_00 , a_=True , a_=True , ): '''simple docstring''' __snake_case : Optional[int] = parent __snake_case : Union[str, Any] = batch_size __snake_case : int = min_seq_length __snake_case : List[str] = max_seq_length __snake_case : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __snake_case : Optional[Any] = feature_size __snake_case : Union[str, Any] = num_mel_bins __snake_case : Tuple = padding_value __snake_case : Optional[Any] = sampling_rate __snake_case : List[str] = return_attention_mask __snake_case : str = do_normalize def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def SCREAMING_SNAKE_CASE (self , a_=False , a_=False ): '''simple docstring''' def _flatten(a_ ): return list(itertools.chain(*a_ ) ) if equal_length: __snake_case : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __snake_case : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __snake_case : Optional[Any] = [np.asarray(a_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _UpperCAmelCase ( __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =SpeechaTextFeatureExtractor if is_speech_available() else None def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = SpeechaTextFeatureExtractionTester(self ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' self.assertTrue(np.all(np.mean(a_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(a_ , axis=0 ) - 1 ) < 1E-3 ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __snake_case : str = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __snake_case : int = [np.asarray(a_ ) for speech_input in speech_inputs] # Test feature size __snake_case : Optional[int] = feature_extractor(a_ , padding=a_ , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __snake_case : str = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __snake_case : Dict = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(a_ , a_ , atol=1E-3 ) ) # Test batched __snake_case : List[str] = feature_extractor(a_ , return_tensors='''np''' ).input_features __snake_case : Optional[int] = feature_extractor(a_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(a_ , a_ ): self.assertTrue(np.allclose(a_ , a_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __snake_case : Tuple = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __snake_case : int = np.asarray(a_ ) __snake_case : str = feature_extractor(a_ , return_tensors='''np''' ).input_features __snake_case : Union[str, Any] = feature_extractor(a_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(a_ , a_ ): self.assertTrue(np.allclose(a_ , a_ , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __snake_case : List[str] = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __snake_case : Union[str, Any] = ['''longest''', '''max_length''', '''do_not_pad'''] __snake_case : int = [None, 16, None] for max_length, padding in zip(a_ , a_ ): __snake_case : List[str] = feature_extractor( a_ , padding=a_ , max_length=a_ , return_attention_mask=a_ ) __snake_case : Union[str, Any] = inputs.input_features __snake_case : Optional[int] = inputs.attention_mask __snake_case : Dict = [np.sum(a_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __snake_case : Optional[int] = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __snake_case : int = ['''longest''', '''max_length''', '''do_not_pad'''] __snake_case : Any = [None, 16, None] for max_length, padding in zip(a_ , a_ ): __snake_case : Dict = feature_extractor( a_ , max_length=a_ , padding=a_ , return_tensors='''np''' , return_attention_mask=a_ ) __snake_case : str = inputs.input_features __snake_case : Any = inputs.attention_mask __snake_case : Union[str, Any] = [np.sum(a_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __snake_case : Union[str, Any] = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __snake_case : str = feature_extractor( a_ , padding='''max_length''' , max_length=4 , truncation=a_ , return_tensors='''np''' , return_attention_mask=a_ , ) __snake_case : Union[str, Any] = inputs.input_features __snake_case : Any = inputs.attention_mask __snake_case : Dict = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __snake_case : int = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __snake_case : Optional[Any] = feature_extractor( a_ , padding='''longest''' , max_length=4 , truncation=a_ , return_tensors='''np''' , return_attention_mask=a_ , ) __snake_case : Optional[Any] = inputs.input_features __snake_case : Union[str, Any] = inputs.attention_mask __snake_case : Union[str, Any] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __snake_case : List[str] = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __snake_case : List[str] = feature_extractor( a_ , padding='''longest''' , max_length=16 , truncation=a_ , return_tensors='''np''' , return_attention_mask=a_ , ) __snake_case : Optional[Any] = inputs.input_features __snake_case : Dict = inputs.attention_mask __snake_case : Tuple = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' import torch __snake_case : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __snake_case : Tuple = np.random.rand(1_00 , 32 ).astype(np.floataa ) __snake_case : int = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __snake_case : Dict = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __snake_case : str = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' from datasets import load_dataset __snake_case : Union[str, Any] = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __snake_case : int = ds.sort('''id''' ).select(range(a_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on __snake_case : Tuple = self._load_datasamples(1 ) __snake_case : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __snake_case : List[str] = feature_extractor(a_ , return_tensors='''pt''' ).input_features self.assertEquals(input_features.shape , (1, 5_84, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , a_ , atol=1E-4 ) )
102
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case : Optional[Any] = {'configuration_xglm': ['XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XGLMConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : int = ['XGLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : List[Any] = ['XGLMTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[Any] = [ 'XGLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XGLMForCausalLM', 'XGLMModel', 'XGLMPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : str = [ 'FlaxXGLMForCausalLM', 'FlaxXGLMModel', 'FlaxXGLMPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[Any] = [ 'TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXGLMForCausalLM', 'TFXGLMModel', 'TFXGLMPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure)
134
0
'''simple docstring''' from random import randint from tempfile import TemporaryFile import numpy as np def UpperCAmelCase ( a_ , a_ , a_ ) -> Dict: """simple docstring""" A_ : List[str] = 0 if start < end: A_ : List[Any] = randint(a_ , a_ ) A_ : Union[str, Any] = a[end] A_ : List[Any] = a[pivot] A_ : str = temp A_ , A_ : int = _in_place_partition(a_ , a_ , a_ ) count += _in_place_quick_sort(a_ , a_ , p - 1 ) count += _in_place_quick_sort(a_ , p + 1 , a_ ) return count def UpperCAmelCase ( a_ , a_ , a_ ) -> List[str]: """simple docstring""" A_ : str = 0 A_ : int = randint(a_ , a_ ) A_ : Optional[Any] = a[end] A_ : Any = a[pivot] A_ : str = temp A_ : Any = start - 1 for index in range(a_ , a_ ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value A_ : Optional[int] = new_pivot_index + 1 A_ : Optional[Any] = a[new_pivot_index] A_ : Union[str, Any] = a[index] A_ : str = temp A_ : int = a[new_pivot_index + 1] A_ : Optional[int] = a[end] A_ : List[Any] = temp return new_pivot_index + 1, count UpperCamelCase__ : Tuple = TemporaryFile() UpperCamelCase__ : Union[str, Any] = 100 # 1000 elements are to be sorted UpperCamelCase__ , UpperCamelCase__ : Dict = 0, 1 # mean and standard deviation UpperCamelCase__ : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('The array is') print(X) outfile.seek(0) # using the same array UpperCamelCase__ : str = np.load(outfile) UpperCamelCase__ : List[str] = len(M) - 1 UpperCamelCase__ : Optional[int] = _in_place_quick_sort(M, 0, r) print( 'No of Comparisons for 100 elements selected from a standard normal distribution' 'is :' ) print(z)
164
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> int: """simple docstring""" A_ : int = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): A_ : Tuple = n - k # Calculate C(n,k) for i in range(a_ ): result *= n - i result //= i + 1 return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , a_ ) // (node_count + 1) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" if n < 0: raise ValueError("""factorial() not defined for negative values""" ) A_ : Union[str, Any] = 1 for i in range(1 , n + 1 ): result *= i return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return catalan_number(a_ ) * factorial(a_ ) if __name__ == "__main__": UpperCamelCase__ : Any = int(input('Enter the number of nodes: ').strip() or 0) if node_count <= 0: raise ValueError('We need some nodes to work with.') print( f'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' f'binary trees and {catalan_number(node_count)} binary search trees.' )
164
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging snake_case__ : List[Any] = logging.get_logger(__name__) class snake_case_( a__ ): __UpperCamelCase = ['''pixel_values'''] def __init__( self : Tuple , UpperCamelCase_ : bool = True , UpperCamelCase_ : Optional[Dict[str, int]] = None , UpperCamelCase_ : PILImageResampling = PILImageResampling.BILINEAR , UpperCamelCase_ : bool = True , UpperCamelCase_ : Dict[str, int] = None , UpperCamelCase_ : bool = True , UpperCamelCase_ : Union[int, float] = 1 / 2_5_5 , UpperCamelCase_ : bool = True , UpperCamelCase_ : Optional[Union[float, List[float]]] = None , UpperCamelCase_ : Optional[Union[float, List[float]]] = None , **UpperCamelCase_ : int , ): super().__init__(**UpperCamelCase_ ) lowerCAmelCase : Any = size if size is not None else {'''shortest_edge''': 2_5_6} lowerCAmelCase : List[str] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) lowerCAmelCase : Dict = crop_size if crop_size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} lowerCAmelCase : List[Any] = get_size_dict(UpperCamelCase_ ) lowerCAmelCase : Optional[Any] = do_resize lowerCAmelCase : List[Any] = size lowerCAmelCase : Optional[Any] = resample lowerCAmelCase : List[str] = do_center_crop lowerCAmelCase : Dict = crop_size lowerCAmelCase : Optional[Any] = do_rescale lowerCAmelCase : List[str] = rescale_factor lowerCAmelCase : List[Any] = do_normalize lowerCAmelCase : Union[str, Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCAmelCase : List[str] = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCamelCase__ ( self : Dict , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : Dict[str, int] , UpperCamelCase_ : PILImageResampling = PILImageResampling.BICUBIC , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : int , ): lowerCAmelCase : List[str] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) lowerCAmelCase : List[str] = get_resize_output_image_size(UpperCamelCase_ , size=size['''shortest_edge'''] , default_to_square=UpperCamelCase_ ) return resize(UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : Tuple , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : Dict[str, int] , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : int , ): lowerCAmelCase : str = get_size_dict(UpperCamelCase_ ) return center_crop(UpperCamelCase_ , size=(size['''height'''], size['''width''']) , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : str , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : float , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : List[str] ): return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : Optional[Any] , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : Union[float, List[float]] , UpperCamelCase_ : Union[float, List[float]] , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : int , ): return normalize(UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : Union[str, Any] , UpperCamelCase_ : ImageInput , UpperCamelCase_ : Optional[bool] = None , UpperCamelCase_ : Dict[str, int] = None , UpperCamelCase_ : PILImageResampling = None , UpperCamelCase_ : bool = None , UpperCamelCase_ : Dict[str, int] = None , UpperCamelCase_ : Optional[bool] = None , UpperCamelCase_ : Optional[float] = None , UpperCamelCase_ : Optional[bool] = None , UpperCamelCase_ : Optional[Union[float, List[float]]] = None , UpperCamelCase_ : Optional[Union[float, List[float]]] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , UpperCamelCase_ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **UpperCamelCase_ : int , ): lowerCAmelCase : List[Any] = do_resize if do_resize is not None else self.do_resize lowerCAmelCase : Dict = size if size is not None else self.size lowerCAmelCase : List[str] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) lowerCAmelCase : Dict = resample if resample is not None else self.resample lowerCAmelCase : int = do_center_crop if do_center_crop is not None else self.do_center_crop lowerCAmelCase : List[str] = crop_size if crop_size is not None else self.crop_size lowerCAmelCase : List[str] = get_size_dict(UpperCamelCase_ ) lowerCAmelCase : List[str] = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase : List[str] = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase : Union[str, Any] = image_mean if image_mean is not None else self.image_mean lowerCAmelCase : str = image_std if image_std is not None else self.image_std lowerCAmelCase : Union[str, Any] = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. lowerCAmelCase : str = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_resize: lowerCAmelCase : Any = [self.resize(image=UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ ) for image in images] if do_center_crop: lowerCAmelCase : int = [self.center_crop(image=UpperCamelCase_ , size=UpperCamelCase_ ) for image in images] if do_rescale: lowerCAmelCase : Optional[Any] = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_normalize: lowerCAmelCase : Tuple = [self.normalize(image=UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ ) for image in images] lowerCAmelCase : Union[str, Any] = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] lowerCAmelCase : Union[str, Any] = {'''pixel_values''': images} return BatchFeature(data=UpperCamelCase_ , tensor_type=UpperCamelCase_ )
60
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_fnet import FNetTokenizer else: snake_case__ : str = None snake_case__ : Optional[Any] = logging.get_logger(__name__) snake_case__ : Optional[int] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} snake_case__ : Dict = { '''vocab_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/spiece.model''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/tokenizer.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/tokenizer.json''', }, } snake_case__ : Any = { '''google/fnet-base''': 512, '''google/fnet-large''': 512, } snake_case__ : Dict = '''▁''' class snake_case_( a__ ): __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = ['''input_ids''', '''token_type_ids'''] __UpperCamelCase = FNetTokenizer def __init__( self : Union[str, Any] , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Any=False , UpperCamelCase_ : Any=True , UpperCamelCase_ : Dict=True , UpperCamelCase_ : Tuple="<unk>" , UpperCamelCase_ : List[str]="[SEP]" , UpperCamelCase_ : List[Any]="<pad>" , UpperCamelCase_ : Union[str, Any]="[CLS]" , UpperCamelCase_ : int="[MASK]" , **UpperCamelCase_ : Optional[Any] , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowerCAmelCase : int = ( AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ , normalized=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else mask_token ) super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , do_lower_case=UpperCamelCase_ , remove_space=UpperCamelCase_ , keep_accents=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , **UpperCamelCase_ , ) lowerCAmelCase : Optional[int] = do_lower_case lowerCAmelCase : str = remove_space lowerCAmelCase : Any = keep_accents lowerCAmelCase : int = vocab_file lowerCAmelCase : List[str] = False if not self.vocab_file else True def lowerCamelCase__ ( self : List[Any] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None ): lowerCAmelCase : Optional[int] = [self.sep_token_id] lowerCAmelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase__ ( self : List[str] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None ): lowerCAmelCase : List[str] = [self.sep_token_id] lowerCAmelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self : List[Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase : str = os.path.join( UpperCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ): copyfile(self.vocab_file , UpperCamelCase_ ) return (out_vocab_file,)
60
1
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCAmelCase_ = logging.get_logger(__name__) @add_end_docstrings(__lowerCamelCase ) class lowercase__ ( __lowerCamelCase ): '''simple docstring''' def __init__( self, **__magic_name__ ) -> Optional[int]: """simple docstring""" super().__init__(**__magic_name__ ) requires_backends(self, '''vision''' ) requires_backends(self, '''torch''' ) if self.framework != "pt": raise ValueError(f"The {self.__class__} is only available in PyTorch." ) self.check_model_type(__magic_name__ ) def UpperCamelCase__ ( self, **__magic_name__ ) -> str: """simple docstring""" UpperCamelCase__ : List[str] = {} UpperCamelCase__ : List[Any] = {} UpperCamelCase__ : Union[str, Any] = {} # preprocess args if "points_per_batch" in kwargs: UpperCamelCase__ : str = kwargs['''points_per_batch'''] if "points_per_crop" in kwargs: UpperCamelCase__ : Optional[int] = kwargs['''points_per_crop'''] if "crops_n_layers" in kwargs: UpperCamelCase__ : Optional[int] = kwargs['''crops_n_layers'''] if "crop_overlap_ratio" in kwargs: UpperCamelCase__ : Any = kwargs['''crop_overlap_ratio'''] if "crop_n_points_downscale_factor" in kwargs: UpperCamelCase__ : Union[str, Any] = kwargs['''crop_n_points_downscale_factor'''] # postprocess args if "pred_iou_thresh" in kwargs: UpperCamelCase__ : Tuple = kwargs['''pred_iou_thresh'''] if "stability_score_offset" in kwargs: UpperCamelCase__ : str = kwargs['''stability_score_offset'''] if "mask_threshold" in kwargs: UpperCamelCase__ : str = kwargs['''mask_threshold'''] if "stability_score_thresh" in kwargs: UpperCamelCase__ : str = kwargs['''stability_score_thresh'''] if "crops_nms_thresh" in kwargs: UpperCamelCase__ : int = kwargs['''crops_nms_thresh'''] if "output_rle_mask" in kwargs: UpperCamelCase__ : int = kwargs['''output_rle_mask'''] if "output_bboxes_mask" in kwargs: UpperCamelCase__ : int = kwargs['''output_bboxes_mask'''] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self, __magic_name__, *__magic_name__, __magic_name__=None, __magic_name__=None, **__magic_name__ ) -> Optional[int]: """simple docstring""" return super().__call__(__magic_name__, *__magic_name__, num_workers=__magic_name__, batch_size=__magic_name__, **__magic_name__ ) def UpperCamelCase__ ( self, __magic_name__, __magic_name__=64, __magic_name__ = 0, __magic_name__ = 512 / 1500, __magic_name__ = 32, __magic_name__ = 1, ) -> List[Any]: """simple docstring""" UpperCamelCase__ : Dict = load_image(__magic_name__ ) UpperCamelCase__ : Tuple = self.image_processor.size['''longest_edge'''] UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ : int = self.image_processor.generate_crop_boxes( __magic_name__, __magic_name__, __magic_name__, __magic_name__, __magic_name__, __magic_name__ ) UpperCamelCase__ : List[str] = self.image_processor(images=__magic_name__, return_tensors='''pt''' ) with self.device_placement(): if self.framework == "pt": UpperCamelCase__ : Tuple = self.get_inference_context() with inference_context(): UpperCamelCase__ : Union[str, Any] = self._ensure_tensor_on_device(__magic_name__, device=self.device ) UpperCamelCase__ : List[Any] = self.model.get_image_embeddings(model_inputs.pop('''pixel_values''' ) ) UpperCamelCase__ : int = image_embeddings UpperCamelCase__ : Optional[Any] = grid_points.shape[1] UpperCamelCase__ : str = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( '''Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. ''' '''To return all points at once, set points_per_batch to None''' ) for i in range(0, __magic_name__, __magic_name__ ): UpperCamelCase__ : Any = grid_points[:, i : i + points_per_batch, :, :] UpperCamelCase__ : str = input_labels[:, i : i + points_per_batch] UpperCamelCase__ : Optional[Any] = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCamelCase__ ( self, __magic_name__, __magic_name__=0.88, __magic_name__=0.95, __magic_name__=0, __magic_name__=1, ) -> Union[str, Any]: """simple docstring""" UpperCamelCase__ : Optional[int] = model_inputs.pop('''input_boxes''' ) UpperCamelCase__ : str = model_inputs.pop('''is_last''' ) UpperCamelCase__ : Any = model_inputs.pop('''original_sizes''' ).tolist() UpperCamelCase__ : List[str] = model_inputs.pop('''reshaped_input_sizes''' ).tolist() UpperCamelCase__ : Optional[Any] = self.model(**__magic_name__ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks UpperCamelCase__ : Any = model_outputs['''pred_masks'''] UpperCamelCase__ : List[str] = self.image_processor.post_process_masks( __magic_name__, __magic_name__, __magic_name__, __magic_name__, binarize=__magic_name__ ) UpperCamelCase__ : int = model_outputs['''iou_scores'''] UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ : Optional[int] = self.image_processor.filter_masks( masks[0], iou_scores[0], original_sizes[0], input_boxes[0], __magic_name__, __magic_name__, __magic_name__, __magic_name__, ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCamelCase__ ( self, __magic_name__, __magic_name__=False, __magic_name__=False, __magic_name__=0.7, ) -> Any: """simple docstring""" UpperCamelCase__ : Optional[Any] = [] UpperCamelCase__ : List[Any] = [] UpperCamelCase__ : Tuple = [] for model_output in model_outputs: all_scores.append(model_output.pop('''iou_scores''' ) ) all_masks.extend(model_output.pop('''masks''' ) ) all_boxes.append(model_output.pop('''boxes''' ) ) UpperCamelCase__ : Union[str, Any] = torch.cat(__magic_name__ ) UpperCamelCase__ : Any = torch.cat(__magic_name__ ) UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ : Union[str, Any] = self.image_processor.post_process_for_mask_generation( __magic_name__, __magic_name__, __magic_name__, __magic_name__ ) UpperCamelCase__ : int = defaultdict(__magic_name__ ) for output in model_outputs: for k, v in output.items(): extra[k].append(__magic_name__ ) UpperCamelCase__ : Optional[int] = {} if output_rle_mask: UpperCamelCase__ : str = rle_mask if output_bboxes_mask: UpperCamelCase__ : Union[str, Any] = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
247
def lowerCAmelCase_ ( __UpperCAmelCase: int ) -> list[int]: if length <= 0 or not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise ValueError('''Length must be a positive integer.''' ) return [n * (2 * n - 1) for n in range(__UpperCAmelCase )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
247
1
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class __snake_case ( unittest.TestCase ): @slow def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = FlaxMTaForConditionalGeneration.from_pretrained("""google/mt5-small""" ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained("""google/mt5-small""" ) SCREAMING_SNAKE_CASE__ = tokenizer("""Hello there""" , return_tensors="""np""" ).input_ids SCREAMING_SNAKE_CASE__ = tokenizer("""Hi I am""" , return_tensors="""np""" ).input_ids SCREAMING_SNAKE_CASE__ = shift_tokens_right(_lowercase , model.config.pad_token_id , model.config.decoder_start_token_id ) SCREAMING_SNAKE_CASE__ = model(_lowercase , decoder_input_ids=_lowercase ).logits SCREAMING_SNAKE_CASE__ = optax.softmax_cross_entropy(_lowercase , onehot(_lowercase , logits.shape[-1] ) ).mean() SCREAMING_SNAKE_CASE__ = -(labels.shape[-1] * loss.item()) SCREAMING_SNAKE_CASE__ = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
219
import requests from bsa import BeautifulSoup def A ( _UpperCAmelCase : str , _UpperCAmelCase : dict ) -> str: '''simple docstring''' _UpperCAmelCase = BeautifulSoup(requests.get(_UpperCAmelCase , params=_UpperCAmelCase ).content , 'html.parser' ) _UpperCAmelCase = soup.find('div' , attrs={'class': 'gs_ri'} ) _UpperCAmelCase = div.find('div' , attrs={'class': 'gs_fl'} ).find_all('a' ) return anchors[2].get_text() if __name__ == "__main__": UpperCAmelCase__ = { "title": ( "Precisely geometry controlled microsupercapacitors for ultrahigh areal " "capacitance, volumetric capacitance, and energy density" ), "journal": "Chem. Mater.", "volume": 30, "pages": "3979-3990", "year": 2018, "hl": "en", } print(get_citation("https://scholar.google.com/scholar_lookup", params=params))
339
0
def __lowercase ( __lowerCAmelCase : int ): if num <= 0: raise ValueError('Input must be a positive integer' ) a__ = [True] * (num + 1) a__ = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowerCAmelCase ): a__ = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() snake_case : Optional[Any] = int(input('''Enter a positive integer: ''').strip()) print(prime_sieve_eratosthenes(user_num))
109
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() snake_case : str = logging.get_logger(__name__) snake_case : Optional[Any] = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn.grep_linear''': '''encoder.layers.*.attention.gru_rel_pos_linear''', '''self_attn.relative_attention_bias''': '''encoder.layers.*.attention.rel_attn_embed''', '''self_attn.grep_a''': '''encoder.layers.*.attention.gru_rel_pos_const''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''ctc_proj''', '''mask_emb''': '''masked_spec_embed''', } snake_case : Dict = [ '''ctc_proj''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def __lowercase ( __lowerCAmelCase : Dict , __lowerCAmelCase : List[str] , __lowerCAmelCase : str , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Dict ): for attribute in key.split('.' ): a__ = getattr(__lowerCAmelCase , __lowerCAmelCase ) if weight_type is not None: a__ = getattr(__lowerCAmelCase , __lowerCAmelCase ).shape else: a__ = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__ = value elif weight_type == "weight_g": a__ = value elif weight_type == "weight_v": a__ = value elif weight_type == "bias": a__ = value else: a__ = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __lowercase ( __lowerCAmelCase : List[str] , __lowerCAmelCase : List[Any] ): a__ = [] a__ = fairseq_model.state_dict() a__ = hf_model.feature_extractor for name, value in fairseq_dict.items(): a__ = False if "conv_layers" in name: load_conv_layer( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) a__ = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: a__ = True if "*" in mapped_key: a__ = name.split(__lowerCAmelCase )[0].split('.' )[-2] a__ = mapped_key.replace('*' , __lowerCAmelCase ) if "weight_g" in name: a__ = 'weight_g' elif "weight_v" in name: a__ = 'weight_v' elif "bias" in name and "relative_attention_bias" not in name: a__ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj a__ = 'weight' else: a__ = None set_recursively(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) continue if not is_used: unused_weights.append(__lowerCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def __lowercase ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Dict , __lowerCAmelCase : int , __lowerCAmelCase : Optional[int] ): a__ = full_name.split('conv_layers.' )[-1] a__ = name.split('.' ) a__ = int(items[0] ) a__ = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) a__ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) a__ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) a__ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) a__ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__lowerCAmelCase ) @torch.no_grad() def __lowercase ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any , __lowerCAmelCase : List[Any]=None ): # load the pre-trained checkpoints a__ = torch.load(__lowerCAmelCase ) a__ = WavLMConfigOrig(checkpoint['cfg'] ) a__ = WavLMOrig(__lowerCAmelCase ) model.load_state_dict(checkpoint['model'] ) model.eval() if config_path is not None: a__ = WavLMConfig.from_pretrained(__lowerCAmelCase ) else: a__ = WavLMConfig() a__ = WavLMModel(__lowerCAmelCase ) recursively_load_weights(__lowerCAmelCase , __lowerCAmelCase ) hf_wavlm.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": snake_case : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') snake_case : Union[str, Any] = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
109
1
'''simple docstring''' import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter A =True except ImportError: A =False A =logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case_ (_a : Namespace ): return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class _a ( __a ): @staticmethod def A ( lowercase : ArgumentParser ): '''simple docstring''' UpperCAmelCase = parser.add_parser('''add-new-model''' ) add_new_model_parser.add_argument('''--testing''' , action='''store_true''' , help='''If in testing mode.''' ) add_new_model_parser.add_argument('''--testing_file''' , type=lowercase , help='''Configuration file on which to run.''' ) add_new_model_parser.add_argument( '''--path''' , type=lowercase , help='''Path to cookiecutter. Should only be used for testing purposes.''' ) add_new_model_parser.set_defaults(func=lowercase ) def __init__( self : List[Any] , lowercase : bool , lowercase : str , lowercase : Dict=None , *lowercase : Optional[Any] ): '''simple docstring''' UpperCAmelCase = testing UpperCAmelCase = testing_file UpperCAmelCase = path def A ( self : Optional[Any] ): '''simple docstring''' warnings.warn( '''The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ''' '''It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ''' '''checks, you should use `transformers-cli add-new-model-like` instead.''' ) if not _has_cookiecutter: raise ImportError( '''Model creation dependencies are required to use the `add_new_model` command. Install them by running ''' '''the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n''' ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory UpperCAmelCase = [directory for directory in os.listdir() if '''cookiecutter-template-''' == directory[:22]] if len(lowercase ) > 0: raise ValueError( '''Several directories starting with `cookiecutter-template-` in current working directory. ''' '''Please clean your directory by removing all folders starting with `cookiecutter-template-` or ''' '''change your working directory.''' ) UpperCAmelCase = ( Path(lowercase ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) UpperCAmelCase = path_to_transformer_root / '''templates''' / '''adding_a_new_model''' # Execute cookiecutter if not self._testing: cookiecutter(str(lowercase ) ) else: with open(self._testing_file , '''r''' ) as configuration_file: UpperCAmelCase = json.load(lowercase ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=lowercase , extra_context=lowercase , ) UpperCAmelCase = [directory for directory in os.listdir() if '''cookiecutter-template-''' in directory[:22]][0] # Retrieve configuration with open(directory + '''/configuration.json''' , '''r''' ) as configuration_file: UpperCAmelCase = json.load(lowercase ) UpperCAmelCase = configuration['''lowercase_modelname'''] UpperCAmelCase = configuration['''generate_tensorflow_pytorch_and_flax'''] os.remove(f"{directory}/configuration.json" ) UpperCAmelCase = '''PyTorch''' in generate_tensorflow_pytorch_and_flax UpperCAmelCase = '''TensorFlow''' in generate_tensorflow_pytorch_and_flax UpperCAmelCase = '''Flax''' in generate_tensorflow_pytorch_and_flax UpperCAmelCase = f"{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}" os.makedirs(lowercase , exist_ok=lowercase ) os.makedirs(f"{path_to_transformer_root}/tests/models/{lowercase_model_name}" , exist_ok=lowercase ) # Tests require submodules as they have parent imports with open(f"{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py" , '''w''' ): pass shutil.move( f"{directory}/__init__.py" , f"{model_dir}/__init__.py" , ) shutil.move( f"{directory}/configuration_{lowercase_model_name}.py" , f"{model_dir}/configuration_{lowercase_model_name}.py" , ) def remove_copy_lines(lowercase : Union[str, Any] ): with open(lowercase , '''r''' ) as f: UpperCAmelCase = f.readlines() with open(lowercase , '''w''' ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(lowercase ) if output_pytorch: if not self._testing: remove_copy_lines(f"{directory}/modeling_{lowercase_model_name}.py" ) shutil.move( f"{directory}/modeling_{lowercase_model_name}.py" , f"{model_dir}/modeling_{lowercase_model_name}.py" , ) shutil.move( f"{directory}/test_modeling_{lowercase_model_name}.py" , f"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py" , ) else: os.remove(f"{directory}/modeling_{lowercase_model_name}.py" ) os.remove(f"{directory}/test_modeling_{lowercase_model_name}.py" ) if output_tensorflow: if not self._testing: remove_copy_lines(f"{directory}/modeling_tf_{lowercase_model_name}.py" ) shutil.move( f"{directory}/modeling_tf_{lowercase_model_name}.py" , f"{model_dir}/modeling_tf_{lowercase_model_name}.py" , ) shutil.move( f"{directory}/test_modeling_tf_{lowercase_model_name}.py" , f"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py" , ) else: os.remove(f"{directory}/modeling_tf_{lowercase_model_name}.py" ) os.remove(f"{directory}/test_modeling_tf_{lowercase_model_name}.py" ) if output_flax: if not self._testing: remove_copy_lines(f"{directory}/modeling_flax_{lowercase_model_name}.py" ) shutil.move( f"{directory}/modeling_flax_{lowercase_model_name}.py" , f"{model_dir}/modeling_flax_{lowercase_model_name}.py" , ) shutil.move( f"{directory}/test_modeling_flax_{lowercase_model_name}.py" , f"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py" , ) else: os.remove(f"{directory}/modeling_flax_{lowercase_model_name}.py" ) os.remove(f"{directory}/test_modeling_flax_{lowercase_model_name}.py" ) shutil.move( f"{directory}/{lowercase_model_name}.md" , f"{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md" , ) shutil.move( f"{directory}/tokenization_{lowercase_model_name}.py" , f"{model_dir}/tokenization_{lowercase_model_name}.py" , ) shutil.move( f"{directory}/tokenization_fast_{lowercase_model_name}.py" , f"{model_dir}/tokenization_{lowercase_model_name}_fast.py" , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(lowercase : str , lowercase : str , lowercase : List[str] ): # Create temp file UpperCAmelCase , UpperCAmelCase = mkstemp() UpperCAmelCase = False with fdopen(lowercase , '''w''' ) as new_file: with open(lowercase ) as old_file: for line in old_file: new_file.write(lowercase ) if line_to_copy_below in line: UpperCAmelCase = True for line_to_copy in lines_to_copy: new_file.write(lowercase ) if not line_found: raise ValueError(f"Line {line_to_copy_below} was not found in file." ) # Copy the file permissions from the old file to the new file copymode(lowercase , lowercase ) # Remove original file remove(lowercase ) # Move new file move(lowercase , lowercase ) def skip_units(lowercase : List[Any] ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(lowercase : Tuple ): with open(lowercase ) as datafile: UpperCAmelCase = [] UpperCAmelCase = False UpperCAmelCase = False for line in datafile: if "# To replace in: " in line and "##" not in line: UpperCAmelCase = line.split('''"''' )[1] UpperCAmelCase = skip_units(lowercase ) elif "# Below: " in line and "##" not in line: UpperCAmelCase = line.split('''"''' )[1] UpperCAmelCase = skip_units(lowercase ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(lowercase , lowercase , lowercase ) UpperCAmelCase = [] elif "# Replace with" in line and "##" not in line: UpperCAmelCase = [] elif "##" not in line: lines_to_copy.append(lowercase ) remove(lowercase ) replace_in_files(f"{directory}/to_replace_{lowercase_model_name}.py" ) os.rmdir(lowercase )
34
'''simple docstring''' def __lowerCAmelCase ( snake_case__ ): return "".join([hex(snake_case__ )[2:].zfill(2 ).upper() for byte in list(snake_case__ )] ) def __lowerCAmelCase ( snake_case__ ): # Check data validity, following RFC3548 # https://www.ietf.org/rfc/rfc3548.txt if (len(snake_case__ ) % 2) != 0: raise ValueError( "Base16 encoded data is invalid:\nData does not have an even number of hex digits." ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(snake_case__ ) <= set("0123456789ABCDEF" ): raise ValueError( "Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters." ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(snake_case__ ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
298
0
"""simple docstring""" from collections.abc import Sequence def UpperCAmelCase ( UpperCAmelCase = None ) -> int: if nums is None or not nums: raise ValueError('Input sequence should not be empty' ) snake_case_ = nums[0] for i in range(1 , len(UpperCAmelCase ) ): snake_case_ = nums[i] snake_case_ = max(UpperCAmelCase , ans + num , UpperCAmelCase ) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user __UpperCamelCase = int(input('''Enter number of elements : ''').strip()) __UpperCamelCase = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
312
"""simple docstring""" from math import pi def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> float: return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
312
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowercase ( _lowerCamelCase , _lowerCamelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase = StableDiffusionXLImgaImgPipeline UpperCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} UpperCAmelCase = PipelineTesterMixin.required_optional_params - {"""latents"""} UpperCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def _snake_case ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCAmelCase : Any = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") ,up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") ,attention_head_dim=(2, 4) ,use_linear_projection=a_ ,addition_embed_type="""text_time""" ,addition_time_embed_dim=8 ,transformer_layers_per_block=(1, 2) ,projection_class_embeddings_input_dim=80 ,cross_attention_dim=64 ,) _UpperCAmelCase : List[str] = EulerDiscreteScheduler( beta_start=0.0_0085 ,beta_end=0.012 ,steps_offset=1 ,beta_schedule="""scaled_linear""" ,timestep_spacing="""leading""" ,) torch.manual_seed(0 ) _UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=4 ,sample_size=128 ,) torch.manual_seed(0 ) _UpperCAmelCase : List[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_000 ,hidden_act="""gelu""" ,projection_dim=32 ,) _UpperCAmelCase : Union[str, Any] = CLIPTextModel(a_ ) _UpperCAmelCase : str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ,local_files_only=a_ ) _UpperCAmelCase : str = CLIPTextModelWithProjection(a_ ) _UpperCAmelCase : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ,local_files_only=a_ ) _UpperCAmelCase : Optional[int] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """text_encoder_2""": text_encoder_a, """tokenizer_2""": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def _snake_case ( self ,a_ ,a_=0 ) -> int: _UpperCAmelCase : str = floats_tensor((1, 3, 32, 32) ,rng=random.Random(a_ ) ).to(a_ ) _UpperCAmelCase : Dict = image / 2 + 0.5 if str(a_ ).startswith("""mps""" ): _UpperCAmelCase : List[Any] = torch.manual_seed(a_ ) else: _UpperCAmelCase : Optional[Any] = torch.Generator(device=a_ ).manual_seed(a_ ) _UpperCAmelCase : List[str] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 5.0, """output_type""": """numpy""", """strength""": 0.75, } return inputs def _snake_case ( self ) -> Union[str, Any]: _UpperCAmelCase : int = """cpu""" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase : Optional[int] = self.get_dummy_components() _UpperCAmelCase : str = StableDiffusionXLImgaImgPipeline(**a_ ) _UpperCAmelCase : Any = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(a_ ) _UpperCAmelCase : Optional[Any] = sd_pipe(**a_ ).images _UpperCAmelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCAmelCase : List[str] = np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _snake_case ( self ) -> Optional[int]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def _snake_case ( self ) -> int: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def _snake_case ( self ) -> Tuple: pass def _snake_case ( self ) -> int: _UpperCAmelCase : str = self.get_dummy_components() _UpperCAmelCase : Tuple = StableDiffusionXLImgaImgPipeline(**a_ ) _UpperCAmelCase : Any = sd_pipe.to(a_ ) _UpperCAmelCase : Optional[int] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) # forward without prompt embeds _UpperCAmelCase : Any = self.get_dummy_inputs(a_ ) _UpperCAmelCase : List[str] = 3 * ["""this is a negative prompt"""] _UpperCAmelCase : str = negative_prompt _UpperCAmelCase : str = 3 * [inputs["""prompt"""]] _UpperCAmelCase : Any = sd_pipe(**a_ ) _UpperCAmelCase : int = output.images[0, -3:, -3:, -1] # forward with prompt embeds _UpperCAmelCase : Dict = self.get_dummy_inputs(a_ ) _UpperCAmelCase : int = 3 * ["""this is a negative prompt"""] _UpperCAmelCase : Tuple = 3 * [inputs.pop("""prompt""" )] ( ( _UpperCAmelCase ) ,( _UpperCAmelCase ) ,( _UpperCAmelCase ) ,( _UpperCAmelCase ) , ) : List[str] = sd_pipe.encode_prompt(a_ ,negative_prompt=a_ ) _UpperCAmelCase : Any = sd_pipe( **a_ ,prompt_embeds=a_ ,negative_prompt_embeds=a_ ,pooled_prompt_embeds=a_ ,negative_pooled_prompt_embeds=a_ ,) _UpperCAmelCase : Union[str, Any] = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class lowercase ( unittest.TestCase ): """simple docstring""" def _snake_case ( self ) -> Union[str, Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self ,a_ ,a_="cpu" ,a_=torch.floataa ,a_=0 ) -> List[Any]: _UpperCAmelCase : Optional[int] = torch.Generator(device=a_ ).manual_seed(a_ ) _UpperCAmelCase : List[Any] = np.random.RandomState(a_ ).standard_normal((1, 4, 64, 64) ) _UpperCAmelCase : List[Any] = torch.from_numpy(a_ ).to(device=a_ ,dtype=a_ ) _UpperCAmelCase : Optional[Any] = { """prompt""": """a photograph of an astronaut riding a horse""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase : Tuple = DiffusionPipeline.from_pretrained("""stabilityai/stable-diffusion-2-base""" ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : Any = self.get_inputs(a_ ) _UpperCAmelCase : Dict = pipe(**a_ ).images _UpperCAmelCase : List[str] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) _UpperCAmelCase : Dict = np.array([0.4_9493, 0.4_7896, 0.4_0798, 0.5_4214, 0.5_3212, 0.4_8202, 0.4_7656, 0.4_6329, 0.4_8506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
215
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ : Optional[Any] = logging.get_logger(__name__) def snake_case_ ( lowerCAmelCase_ )-> Optional[Any]: '''simple docstring''' _UpperCAmelCase : Dict = DPTConfig() if "large" in checkpoint_url: _UpperCAmelCase : List[Any] = 1024 _UpperCAmelCase : Optional[int] = 4096 _UpperCAmelCase : Tuple = 24 _UpperCAmelCase : List[str] = 16 _UpperCAmelCase : str = [5, 11, 17, 23] _UpperCAmelCase : Tuple = [256, 512, 1024, 1024] _UpperCAmelCase : List[str] = (1, 384, 384) if "ade" in checkpoint_url: _UpperCAmelCase : Optional[int] = True _UpperCAmelCase : Tuple = 150 _UpperCAmelCase : Tuple = """huggingface/label-files""" _UpperCAmelCase : int = """ade20k-id2label.json""" _UpperCAmelCase : List[str] = json.load(open(cached_download(hf_hub_url(lowerCAmelCase_ , lowerCAmelCase_ , repo_type="""dataset""" ) ) , """r""" ) ) _UpperCAmelCase : List[Any] = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} _UpperCAmelCase : Tuple = idalabel _UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} _UpperCAmelCase : Optional[int] = [1, 150, 480, 480] return config, expected_shape def snake_case_ ( lowerCAmelCase_ )-> str: '''simple docstring''' _UpperCAmelCase : Tuple = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(lowerCAmelCase_ , lowerCAmelCase_ ) def snake_case_ ( lowerCAmelCase_ )-> Any: '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): _UpperCAmelCase : int = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: _UpperCAmelCase : str = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: _UpperCAmelCase : Optional[Any] = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: _UpperCAmelCase : int = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: _UpperCAmelCase : Dict = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: _UpperCAmelCase : List[str] = name.replace("""proj""" , """projection""" ) if "blocks" in name: _UpperCAmelCase : Dict = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: _UpperCAmelCase : Optional[Any] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _UpperCAmelCase : Dict = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: _UpperCAmelCase : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _UpperCAmelCase : Dict = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: _UpperCAmelCase : Dict = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: _UpperCAmelCase : Optional[Any] = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: _UpperCAmelCase : List[Any] = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: _UpperCAmelCase : Optional[int] = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: _UpperCAmelCase : List[str] = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: _UpperCAmelCase : str = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: _UpperCAmelCase : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 _UpperCAmelCase : Tuple = name.replace(F'''refinenet{layer_idx}''' , F'''fusion_stage.layers.{abs(layer_idx-4 )}''' ) if "out_conv" in name: _UpperCAmelCase : List[Any] = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: _UpperCAmelCase : Tuple = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: _UpperCAmelCase : Union[str, Any] = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: _UpperCAmelCase : int = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: _UpperCAmelCase : List[str] = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: _UpperCAmelCase : Optional[Any] = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: _UpperCAmelCase : List[str] = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: _UpperCAmelCase : Tuple = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: _UpperCAmelCase : List[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: _UpperCAmelCase : List[str] = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: _UpperCAmelCase : Optional[int] = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: _UpperCAmelCase : int = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: _UpperCAmelCase : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: _UpperCAmelCase : str = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: _UpperCAmelCase : Optional[Any] = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: _UpperCAmelCase : Optional[Any] = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: _UpperCAmelCase : Any = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: _UpperCAmelCase : Tuple = name.replace("""bn""" , """batch_norm""" ) if "head" in name: _UpperCAmelCase : Dict = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: _UpperCAmelCase : List[Any] = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: _UpperCAmelCase : List[Any] = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ )-> Any: '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase : Any = state_dict.pop(F'''dpt.encoder.layer.{i}.attn.qkv.weight''' ) _UpperCAmelCase : Union[str, Any] = state_dict.pop(F'''dpt.encoder.layer.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase : Optional[int] = in_proj_weight[: config.hidden_size, :] _UpperCAmelCase : Optional[Any] = in_proj_bias[: config.hidden_size] _UpperCAmelCase : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase : Tuple = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase : Dict = in_proj_bias[-config.hidden_size :] def snake_case_ ( )-> Optional[int]: '''simple docstring''' _UpperCAmelCase : Dict = """http://images.cocodataset.org/val2017/000000039769.jpg""" _UpperCAmelCase : List[str] = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> Dict: '''simple docstring''' _UpperCAmelCase ,_UpperCAmelCase : List[str] = get_dpt_config(lowerCAmelCase_ ) # load original state_dict from URL _UpperCAmelCase : Any = torch.hub.load_state_dict_from_url(lowerCAmelCase_ , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(lowerCAmelCase_ ) # rename keys for key in state_dict.copy().keys(): _UpperCAmelCase : List[Any] = state_dict.pop(lowerCAmelCase_ ) _UpperCAmelCase : Dict = val # read in qkv matrices read_in_q_k_v(lowerCAmelCase_ , lowerCAmelCase_ ) # load HuggingFace model _UpperCAmelCase : Optional[int] = DPTForSemanticSegmentation(lowerCAmelCase_ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(lowerCAmelCase_ ) model.load_state_dict(lowerCAmelCase_ ) model.eval() # Check outputs on an image _UpperCAmelCase : Tuple = 480 if """ade""" in checkpoint_url else 384 _UpperCAmelCase : List[str] = DPTImageProcessor(size=lowerCAmelCase_ ) _UpperCAmelCase : Optional[Any] = prepare_img() _UpperCAmelCase : Dict = image_processor(lowerCAmelCase_ , return_tensors="""pt""" ) # forward pass _UpperCAmelCase : Optional[Any] = model(**lowerCAmelCase_ ).logits if """ade""" in checkpoint_url else model(**lowerCAmelCase_ ).predicted_depth # Assert logits _UpperCAmelCase : Optional[int] = torch.tensor([[6.3_1_9_9, 6.3_6_2_9, 6.4_1_4_8], [6.3_8_5_0, 6.3_6_1_5, 6.4_1_6_6], [6.3_5_1_9, 6.3_1_7_6, 6.3_5_7_5]] ) if "ade" in checkpoint_url: _UpperCAmelCase : str = torch.tensor([[4.0_4_8_0, 4.2_4_2_0, 4.4_3_6_0], [4.3_1_2_4, 4.5_6_9_3, 4.8_2_6_1], [4.5_7_6_8, 4.8_9_6_5, 5.2_1_6_3]] ) assert outputs.shape == torch.Size(lowerCAmelCase_ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , lowerCAmelCase_ , atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , lowerCAmelCase_ ) ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase_ ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase_ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(lowerCAmelCase_ , lowerCAmelCase_ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=lowerCAmelCase_ , ) image_processor.push_to_hub( repo_path_or_name=Path(lowerCAmelCase_ , lowerCAmelCase_ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=lowerCAmelCase_ , ) if __name__ == "__main__": A_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt""", type=str, help="""URL of the original DPT checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) parser.add_argument( """--model_name""", default="""dpt-large""", type=str, help="""Name of the model, in case you're pushing to the hub.""", ) A_ : List[Any] = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
215
1
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowercase__ ='''▁''' lowercase__ =get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class UpperCamelCase__ ( a__ ,unittest.TestCase ): _SCREAMING_SNAKE_CASE : Dict = BertGenerationTokenizer _SCREAMING_SNAKE_CASE : int = False _SCREAMING_SNAKE_CASE : List[str] = True def lowerCAmelCase (self : List[str] ): super().setUp() __a : Optional[int] = BertGenerationTokenizer(_lowerCamelCase , keep_accents=_lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase (self : int ): __a : List[str] = '''<s>''' __a : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCamelCase ) , _lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCamelCase ) , _lowerCamelCase ) def lowerCAmelCase (self : List[str] ): __a : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''<pad>''' ) self.assertEqual(len(_lowerCamelCase ) , 1_0_0_2 ) def lowerCAmelCase (self : Union[str, Any] ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0 ) def lowerCAmelCase (self : int ): __a : Dict = BertGenerationTokenizer(_lowerCamelCase , keep_accents=_lowerCamelCase ) __a : Optional[int] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowerCamelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2] , ) __a : str = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __a : Optional[Any] = tokenizer.convert_tokens_to_ids(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4] , ) __a : Any = tokenizer.convert_ids_to_tokens(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase (self : Tuple ): return BertGenerationTokenizer.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) @slow def lowerCAmelCase (self : str ): __a : List[Any] = '''Hello World!''' __a : Union[str, Any] = [1_8_5_3_6, 2_2_6_0, 1_0_1] self.assertListEqual(_lowerCamelCase , self.big_tokenizer.encode(_lowerCamelCase ) ) @slow def lowerCAmelCase (self : Tuple ): __a : List[str] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) __a : List[str] = [ 8_7_1, 4_1_9, 3_5_8, 9_4_6, 9_9_1, 2_5_2_1, 4_5_2, 3_5_8, 1_3_5_7, 3_8_7, 7_7_5_1, 3_5_3_6, 1_1_2, 9_8_5, 4_5_6, 1_2_6, 8_6_5, 9_3_8, 5_4_0_0, 5_7_3_4, 4_5_8, 1_3_6_8, 4_6_7, 7_8_6, 2_4_6_2, 5_2_4_6, 1_1_5_9, 6_3_3, 8_6_5, 4_5_1_9, 4_5_7, 5_8_2, 8_5_2, 2_5_5_7, 4_2_7, 9_1_6, 5_0_8, 4_0_5, 3_4_3_2_4, 4_9_7, 3_9_1, 4_0_8, 1_1_3_4_2, 1_2_4_4, 3_8_5, 1_0_0, 9_3_8, 9_8_5, 4_5_6, 5_7_4, 3_6_2, 1_2_5_9_7, 3_2_0_0, 3_1_2_9, 1_1_7_2, ] self.assertListEqual(_lowerCamelCase , self.big_tokenizer.encode(_lowerCamelCase ) ) @require_torch @slow def lowerCAmelCase (self : List[str] ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence __a : List[str] = list(self.big_tokenizer.get_vocab().keys() )[:1_0] __a : str = ''' '''.join(_lowerCamelCase ) __a : Dict = self.big_tokenizer.encode_plus(_lowerCamelCase , return_tensors='''pt''' , return_token_type_ids=_lowerCamelCase ) __a : Dict = self.big_tokenizer.batch_encode_plus( [sequence + ''' ''' + sequence] , return_tensors='''pt''' , return_token_type_ids=_lowerCamelCase ) __a : Dict = BertGenerationConfig() __a : str = BertGenerationEncoder(_lowerCamelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_lowerCamelCase ) model(**_lowerCamelCase ) @slow def lowerCAmelCase (self : str ): # fmt: off __a : int = {'''input_ids''': [[3_9_2_8_6, 4_5_8, 3_6_3_3_5, 2_0_0_1, 4_5_6, 1_3_0_7_3, 1_3_2_6_6, 4_5_5, 1_1_3, 7_7_4_6, 1_7_4_1, 1_1_1_5_7, 3_9_1, 1_3_0_7_3, 1_3_2_6_6, 4_5_5, 1_1_3, 3_9_6_7, 3_5_4_1_2, 1_1_3, 4_9_3_6, 1_0_9, 3_8_7_0, 2_3_7_7, 1_1_3, 3_0_0_8_4, 4_5_7_2_0, 4_5_8, 1_3_4, 1_7_4_9_6, 1_1_2, 5_0_3, 1_1_6_7_2, 1_1_3, 1_1_8, 1_1_2, 5_6_6_5, 1_3_3_4_7, 3_8_6_8_7, 1_1_2, 1_4_9_6, 3_1_3_8_9, 1_1_2, 3_2_6_8, 4_7_2_6_4, 1_3_4, 9_6_2, 1_1_2, 1_6_3_7_7, 8_0_3_5, 2_3_1_3_0, 4_3_0, 1_2_1_6_9, 1_5_5_1_8, 2_8_5_9_2, 4_5_8, 1_4_6, 4_1_6_9_7, 1_0_9, 3_9_1, 1_2_1_6_9, 1_5_5_1_8, 1_6_6_8_9, 4_5_8, 1_4_6, 4_1_3_5_8, 1_0_9, 4_5_2, 7_2_6, 4_0_3_4, 1_1_1, 7_6_3, 3_5_4_1_2, 5_0_8_2, 3_8_8, 1_9_0_3, 1_1_1, 9_0_5_1, 3_9_1, 2_8_7_0, 4_8_9_1_8, 1_9_0_0, 1_1_2_3, 5_5_0, 9_9_8, 1_1_2, 9_5_8_6, 1_5_9_8_5, 4_5_5, 3_9_1, 4_1_0, 2_2_9_5_5, 3_7_6_3_6, 1_1_4], [4_4_8, 1_7_4_9_6, 4_1_9, 3_6_6_3, 3_8_5, 7_6_3, 1_1_3, 2_7_5_3_3, 2_8_7_0, 3_2_8_3, 1_3_0_4_3, 1_6_3_9, 2_4_7_1_3, 5_2_3, 6_5_6, 2_4_0_1_3, 1_8_5_5_0, 2_5_2_1, 5_1_7, 2_7_0_1_4, 2_1_2_4_4, 4_2_0, 1_2_1_2, 1_4_6_5, 3_9_1, 9_2_7, 4_8_3_3, 3_8_8, 5_7_8, 1_1_7_8_6, 1_1_4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_8_4, 2_1_6_9, 7_6_8_7, 2_1_9_3_2, 1_8_1_4_6, 7_2_6, 3_6_3, 1_7_0_3_2, 3_3_9_1, 1_1_4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCamelCase , model_name='''google/bert_for_seq_generation_L-24_bbc_encoder''' , revision='''c817d1fd1be2ffa69431227a1fe320544943d4db''' , )
369
def __UpperCamelCase ( lowerCAmelCase__ : list[list[int | float]] ): __a : int = len(lowerCAmelCase__ ) __a : Dict = len(matrix[0] ) __a : Union[str, Any] = min(lowerCAmelCase__ , lowerCAmelCase__ ) for row in range(lowerCAmelCase__ ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , lowerCAmelCase__ ): __a : Dict = matrix[col][row] / matrix[row][row] for i in range(lowerCAmelCase__ , lowerCAmelCase__ ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __a : Optional[int] = True for i in range(row + 1 , lowerCAmelCase__ ): if matrix[i][row] != 0: __a , __a : Any = matrix[i], matrix[row] __a : Union[str, Any] = False break if reduce: rank -= 1 for i in range(lowerCAmelCase__ ): __a : Optional[Any] = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
90
0
import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def lowerCamelCase_ ( UpperCamelCase__ : Union[str, Any] ) -> Any: """simple docstring""" __lowerCamelCase = [] if isinstance(_lowerCamelCase , _lowerCamelCase ): for v in tree.values(): shapes.extend(_fetch_dims(_lowerCamelCase ) ) elif isinstance(_lowerCamelCase , (list, tuple) ): for t in tree: shapes.extend(_fetch_dims(_lowerCamelCase ) ) elif isinstance(_lowerCamelCase , torch.Tensor ): shapes.append(tree.shape ) else: raise ValueError('Not supported' ) return shapes @torch.jit.ignore def lowerCamelCase_ ( UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple ) -> str: """simple docstring""" __lowerCamelCase = [] for d in reversed(_lowerCamelCase ): idx.append(flat_idx % d ) __lowerCamelCase = flat_idx // d return tuple(reversed(_lowerCamelCase ) ) @torch.jit.ignore def lowerCamelCase_ ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Dict , UpperCamelCase__ : Any = None , UpperCamelCase__ : Union[str, Any] = None , ) -> Optional[int]: """simple docstring""" def reduce_edge_list(UpperCamelCase__ : Dict ) -> None: __lowerCamelCase = True for i in range(len(_lowerCamelCase ) ): __lowerCamelCase = -1 * (i + 1) l[reversed_idx] &= tally __lowerCamelCase = l[reversed_idx] if start_edges is None: __lowerCamelCase = [s == 0 for s in start] reduce_edge_list(_lowerCamelCase ) if end_edges is None: __lowerCamelCase = [e == (d - 1) for e, d in zip(_lowerCamelCase , _lowerCamelCase )] reduce_edge_list(_lowerCamelCase ) # Base cases. Either start/end are empty and we're done, or the final, # one-dimensional tensor can be simply sliced if len(_lowerCamelCase ) == 0: return [()] elif len(_lowerCamelCase ) == 1: return [(slice(start[0] , end[0] + 1 ),)] __lowerCamelCase = [] __lowerCamelCase = [] # Dimensions common to start and end can be selected directly for s, e in zip(_lowerCamelCase , _lowerCamelCase ): if s == e: path_list.append(slice(_lowerCamelCase , s + 1 ) ) else: break __lowerCamelCase = tuple(_lowerCamelCase ) __lowerCamelCase = len(_lowerCamelCase ) # start == end, and we're done if divergence_idx == len(_lowerCamelCase ): return [path] def upper() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None __lowerCamelCase = start[divergence_idx] return tuple( path + (slice(_lowerCamelCase , sdi + 1 ),) + s for s in _get_minimal_slice_set( start[divergence_idx + 1 :] , [d - 1 for d in dims[divergence_idx + 1 :]] , dims[divergence_idx + 1 :] , start_edges=start_edges[divergence_idx + 1 :] , end_edges=[True for _ in end_edges[divergence_idx + 1 :]] , ) ) def lower() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None __lowerCamelCase = end[divergence_idx] return tuple( path + (slice(_lowerCamelCase , edi + 1 ),) + s for s in _get_minimal_slice_set( [0 for _ in start[divergence_idx + 1 :]] , end[divergence_idx + 1 :] , dims[divergence_idx + 1 :] , start_edges=[True for _ in start_edges[divergence_idx + 1 :]] , end_edges=end_edges[divergence_idx + 1 :] , ) ) # If both start and end are at the edges of the subtree rooted at # divergence_idx, we can just select the whole subtree at once if start_edges[divergence_idx] and end_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] + 1 ),) ) # If just start is at the edge, we can grab almost all of the subtree, # treating only the ragged bottom edge as an edge case elif start_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] ),) ) slices.extend(lower() ) # Analogous to the previous case, but the top is ragged this time elif end_edges[divergence_idx]: slices.extend(upper() ) slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] + 1 ),) ) # If both sides of the range are ragged, we need to handle both sides # separately. If there's contiguous meat in between them, we can index it # in one big chunk else: slices.extend(upper() ) __lowerCamelCase = end[divergence_idx] - start[divergence_idx] if middle_ground > 1: slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] ),) ) slices.extend(lower() ) return slices @torch.jit.ignore def lowerCamelCase_ ( UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Tuple , UpperCamelCase__ : List[str] ) -> Tuple: """simple docstring""" __lowerCamelCase = t.shape[:no_batch_dims] __lowerCamelCase = list(_flat_idx_to_idx(_lowerCamelCase , _lowerCamelCase ) ) # _get_minimal_slice_set is inclusive __lowerCamelCase = list(_flat_idx_to_idx(flat_end - 1 , _lowerCamelCase ) ) # Get an ordered list of slices to perform __lowerCamelCase = _get_minimal_slice_set( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) __lowerCamelCase = [t[s] for s in slices] return torch.cat([s.view((-1,) + t.shape[no_batch_dims:] ) for s in sliced_tensors] ) def lowerCamelCase_ ( UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple , UpperCamelCase__ : List[str] , UpperCamelCase__ : Dict , UpperCamelCase__ : str = False , UpperCamelCase__ : Tuple = None , UpperCamelCase__ : Any = False , ) -> Optional[Any]: """simple docstring""" if not (len(_lowerCamelCase ) > 0): raise ValueError('Must provide at least one input' ) __lowerCamelCase = [shape[:no_batch_dims] for shape in _fetch_dims(_lowerCamelCase )] __lowerCamelCase = tuple([max(_lowerCamelCase ) for s in zip(*_lowerCamelCase )] ) def _prep_inputs(UpperCamelCase__ : str ) -> torch.Tensor: if not low_mem: if not sum(t.shape[:no_batch_dims] ) == no_batch_dims: __lowerCamelCase = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) __lowerCamelCase = t.reshape(-1 , *t.shape[no_batch_dims:] ) else: __lowerCamelCase = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) return t __lowerCamelCase = tensor_tree_map(_prep_inputs , _lowerCamelCase ) __lowerCamelCase = None if _out is not None: __lowerCamelCase = tensor_tree_map(lambda UpperCamelCase__ : t.view([-1] + list(t.shape[no_batch_dims:] ) ) , _out ) __lowerCamelCase = 1 for d in orig_batch_dims: flat_batch_dim *= d __lowerCamelCase = flat_batch_dim // chunk_size + (flat_batch_dim % chunk_size != 0) def _select_chunk(UpperCamelCase__ : List[str] ) -> torch.Tensor: return t[i : i + chunk_size] if t.shape[0] != 1 else t __lowerCamelCase = 0 __lowerCamelCase = prepped_outputs for _ in range(_lowerCamelCase ): # Chunk the input if not low_mem: __lowerCamelCase = _select_chunk else: __lowerCamelCase = partial( _chunk_slice , flat_start=_lowerCamelCase , flat_end=min(_lowerCamelCase , i + chunk_size ) , no_batch_dims=len(_lowerCamelCase ) , ) __lowerCamelCase = tensor_tree_map(_lowerCamelCase , _lowerCamelCase ) # Run the layer on the chunk __lowerCamelCase = layer(**_lowerCamelCase ) # Allocate space for the output if out is None: __lowerCamelCase = tensor_tree_map(lambda UpperCamelCase__ : t.new_zeros((flat_batch_dim,) + t.shape[1:] ) , _lowerCamelCase ) # Put the chunk in its pre-allocated space if isinstance(_lowerCamelCase , _lowerCamelCase ): def assign(UpperCamelCase__ : Any , UpperCamelCase__ : List[str] ) -> None: for k, v in da.items(): if isinstance(_lowerCamelCase , _lowerCamelCase ): assign(_lowerCamelCase , da[k] ) else: if _add_into_out: v[i : i + chunk_size] += da[k] else: __lowerCamelCase = da[k] assign(_lowerCamelCase , _lowerCamelCase ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): for xa, xa in zip(_lowerCamelCase , _lowerCamelCase ): if _add_into_out: xa[i : i + chunk_size] += xa else: __lowerCamelCase = xa elif isinstance(_lowerCamelCase , torch.Tensor ): if _add_into_out: out[i : i + chunk_size] += output_chunk else: __lowerCamelCase = output_chunk else: raise ValueError('Not supported' ) i += chunk_size __lowerCamelCase = tensor_tree_map(lambda UpperCamelCase__ : t.view(orig_batch_dims + t.shape[1:] ) , _lowerCamelCase ) return out class __lowerCAmelCase : """simple docstring""" def __init__( self , lowerCamelCase__ = 512 , ) -> str: '''simple docstring''' __lowerCamelCase = max_chunk_size __lowerCamelCase = None __lowerCamelCase = None def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> str: '''simple docstring''' logging.info('Tuning chunk size...' ) if min_chunk_size >= self.max_chunk_size: return min_chunk_size __lowerCamelCase = [2**l for l in range(int(math.log(self.max_chunk_size , 2 ) ) + 1 )] __lowerCamelCase = [c for c in candidates if c > min_chunk_size] __lowerCamelCase = [min_chunk_size] + candidates candidates[-1] += 4 def test_chunk_size(lowerCamelCase__ ) -> bool: try: with torch.no_grad(): fn(*__a , chunk_size=__a ) return True except RuntimeError: return False __lowerCamelCase = 0 __lowerCamelCase = len(__a ) - 1 while i > min_viable_chunk_size_index: __lowerCamelCase = test_chunk_size(candidates[i] ) if not viable: __lowerCamelCase = (min_viable_chunk_size_index + i) // 2 else: __lowerCamelCase = i __lowerCamelCase = (i + len(__a ) - 1) // 2 return candidates[min_viable_chunk_size_index] def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ ) -> List[Any]: '''simple docstring''' __lowerCamelCase = True for aa, aa in zip(__a , __a ): assert type(__a ) == type(__a ) if isinstance(__a , (list, tuple) ): consistent &= self._compare_arg_caches(__a , __a ) elif isinstance(__a , __a ): __lowerCamelCase = [v for _, v in sorted(aa.items() , key=lambda lowerCamelCase__ : x[0] )] __lowerCamelCase = [v for _, v in sorted(aa.items() , key=lambda lowerCamelCase__ : x[0] )] consistent &= self._compare_arg_caches(__a , __a ) else: consistent &= aa == aa return consistent def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = True __lowerCamelCase = tree_map(lambda lowerCamelCase__ : a.shape if isinstance(__a , torch.Tensor ) else a , __a , __a ) if self.cached_arg_data is not None: # If args have changed shape/value, we need to re-tune assert len(self.cached_arg_data ) == len(__a ) __lowerCamelCase = self._compare_arg_caches(self.cached_arg_data , __a ) else: # Otherwise, we can reuse the precomputed value __lowerCamelCase = False if not consistent: __lowerCamelCase = self._determine_favorable_chunk_size( __a , __a , __a , ) __lowerCamelCase = arg_data assert self.cached_chunk_size is not None return self.cached_chunk_size
90
from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( 'The RoBERTa Model transformer with early exiting (DeeRoBERTa). ' , a , ) class UpperCAmelCase_ ( a): lowerCamelCase__ = RobertaConfig lowerCamelCase__ = 'roberta' def __init__( self, __a): '''simple docstring''' super().__init__(__a) _lowerCAmelCase : Optional[Any] = RobertaEmbeddings(__a) self.init_weights() @add_start_docstrings( 'RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top,\n also takes care of multi-layer training. ' , a , ) class UpperCAmelCase_ ( a): lowerCamelCase__ = RobertaConfig lowerCamelCase__ = 'roberta' def __init__( self, __a): '''simple docstring''' super().__init__(__a) _lowerCAmelCase : Optional[int] = config.num_labels _lowerCAmelCase : Optional[int] = config.num_hidden_layers _lowerCAmelCase : Optional[int] = DeeRobertaModel(__a) _lowerCAmelCase : Union[str, Any] = nn.Dropout(config.hidden_dropout_prob) _lowerCAmelCase : List[str] = nn.Linear(config.hidden_size, self.config.num_labels) @add_start_docstrings_to_model_forward(__a) def snake_case__ ( self, __a=None, __a=None, __a=None, __a=None, __a=None, __a=None, __a=None, __a=-1, __a=False, ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self.num_layers try: _lowerCAmelCase : List[Any] = self.roberta( __a, attention_mask=__a, token_type_ids=__a, position_ids=__a, head_mask=__a, inputs_embeds=__a, ) _lowerCAmelCase : List[Any] = outputs[1] _lowerCAmelCase : Dict = self.dropout(__a) _lowerCAmelCase : Dict = self.classifier(__a) _lowerCAmelCase : Optional[Any] = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: _lowerCAmelCase : Tuple = e.message _lowerCAmelCase : Union[str, Any] = e.exit_layer _lowerCAmelCase : List[Any] = outputs[0] if not self.training: _lowerCAmelCase : int = entropy(__a) _lowerCAmelCase : List[Any] = [] _lowerCAmelCase : str = [] if labels is not None: if self.num_labels == 1: # We are doing regression _lowerCAmelCase : Optional[Any] = MSELoss() _lowerCAmelCase : int = loss_fct(logits.view(-1), labels.view(-1)) else: _lowerCAmelCase : Optional[Any] = CrossEntropyLoss() _lowerCAmelCase : Optional[Any] = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) # work with highway exits _lowerCAmelCase : Optional[int] = [] for highway_exit in outputs[-1]: _lowerCAmelCase : Any = highway_exit[0] if not self.training: highway_logits_all.append(__a) highway_entropy.append(highway_exit[2]) if self.num_labels == 1: # We are doing regression _lowerCAmelCase : List[str] = MSELoss() _lowerCAmelCase : List[Any] = loss_fct(highway_logits.view(-1), labels.view(-1)) else: _lowerCAmelCase : Dict = CrossEntropyLoss() _lowerCAmelCase : Optional[Any] = loss_fct(highway_logits.view(-1, self.num_labels), labels.view(-1)) highway_losses.append(__a) if train_highway: _lowerCAmelCase : int = (sum(highway_losses[:-1]),) + outputs # exclude the final highway, of course else: _lowerCAmelCase : Any = (loss,) + outputs if not self.training: _lowerCAmelCase : Optional[Any] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: _lowerCAmelCase : Optional[Any] = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
36
0
'''simple docstring''' import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def __magic_name__( lowerCamelCase, lowerCamelCase=False): try: __lowerCAmelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __lowerCAmelCase = default else: # KEY is set, convert it to True or False. try: __lowerCAmelCase = strtobool(lowerCamelCase) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F"""If set, {key} must be yes or no.""") return _value _UpperCAmelCase : Union[str, Any] = parse_flag_from_env("""RUN_SLOW""", default=False) def __magic_name__( lowerCamelCase): return unittest.skip('''Test was skipped''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(_run_slow_tests, '''test is slow''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(not torch.cuda.is_available(), '''test requires only a CPU''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(torch.cuda.is_available(), '''test requires a GPU''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_xpu_available(), '''test requires a XPU''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_mps_available(), '''test requires a `mps` backend support in `torch`''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless( is_transformers_available() and is_datasets_available(), '''test requires the Hugging Face suite''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_bnb_available(), '''test requires the bitsandbytes library''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_tpu_available(), '''test requires TPU''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(torch.cuda.device_count() == 1, '''test requires a GPU''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(torch.xpu.device_count() == 1, '''test requires a XPU''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(torch.cuda.device_count() > 1, '''test requires multiple GPUs''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(torch.xpu.device_count() > 1, '''test requires multiple XPUs''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_safetensors_available(), '''test requires safetensors''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_deepspeed_available(), '''test requires DeepSpeed''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_torch_version('''>=''', '''1.12.0'''), '''test requires torch version >= 1.12.0''')(lowerCamelCase) def __magic_name__( lowerCamelCase=None, lowerCamelCase=None): if test_case is None: return partial(lowerCamelCase, version=lowerCamelCase) return unittest.skipUnless(is_torch_version('''>=''', lowerCamelCase), F"""test requires torch version >= {version}""")(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_tensorboard_available(), '''test requires Tensorboard''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_wandb_available(), '''test requires wandb''')(lowerCamelCase) def __magic_name__( lowerCamelCase): return unittest.skipUnless(is_comet_ml_available(), '''test requires comet_ml''')(lowerCamelCase) _UpperCAmelCase : Union[str, Any] = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def __magic_name__( lowerCamelCase): return unittest.skipUnless( _atleast_one_tracker_available, '''test requires at least one tracker to be available and for `comet_ml` to not be installed''', )(lowerCamelCase) class a__ ( unittest.TestCase ): """simple docstring""" __UpperCamelCase : int = True @classmethod def _snake_case (cls ): __lowerCAmelCase = tempfile.mkdtemp() @classmethod def _snake_case (cls ): if os.path.exists(cls.tmpdir ): shutil.rmtree(cls.tmpdir ) def _snake_case (self ): if self.clear_on_setup: for path in Path(self.tmpdir ).glob('''**/*''' ): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(__lowercase ) class a__ ( unittest.TestCase ): """simple docstring""" def _snake_case (self ): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class a__ ( unittest.TestCase ): """simple docstring""" def _snake_case (self , __lowercase ): __lowerCAmelCase = mocks if isinstance(__lowercase , (tuple, list) ) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop ) def __magic_name__( lowerCamelCase): __lowerCAmelCase = AcceleratorState() __lowerCAmelCase = tensor[None].clone().to(state.device) __lowerCAmelCase = gather(lowerCamelCase).cpu() __lowerCAmelCase = tensor[0].cpu() for i in range(tensors.shape[0]): if not torch.equal(tensors[i], lowerCamelCase): return False return True class a__ : """simple docstring""" def __init__(self , __lowercase , __lowercase , __lowercase ): __lowerCAmelCase = returncode __lowerCAmelCase = stdout __lowerCAmelCase = stderr async def __magic_name__( lowerCamelCase, lowerCamelCase): while True: __lowerCAmelCase = await stream.readline() if line: callback(lowerCamelCase) else: break async def __magic_name__( lowerCamelCase, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=False, lowerCamelCase=False): if echo: print('''\nRunning: ''', ''' '''.join(lowerCamelCase)) __lowerCAmelCase = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=lowerCamelCase, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=lowerCamelCase, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __lowerCAmelCase = [] __lowerCAmelCase = [] def tee(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=""): __lowerCAmelCase = line.decode('''utf-8''').rstrip() sink.append(lowerCamelCase) if not quiet: print(lowerCamelCase, lowerCamelCase, file=lowerCamelCase) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout, lambda lowerCamelCase: tee(lowerCamelCase, lowerCamelCase, sys.stdout, label='''stdout:'''))), asyncio.create_task(_read_stream(p.stderr, lambda lowerCamelCase: tee(lowerCamelCase, lowerCamelCase, sys.stderr, label='''stderr:'''))), ], timeout=lowerCamelCase, ) return _RunOutput(await p.wait(), lowerCamelCase, lowerCamelCase) def __magic_name__( lowerCamelCase, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=1_8_0, lowerCamelCase=False, lowerCamelCase=True): __lowerCAmelCase = asyncio.get_event_loop() __lowerCAmelCase = loop.run_until_complete( _stream_subprocess(lowerCamelCase, env=lowerCamelCase, stdin=lowerCamelCase, timeout=lowerCamelCase, quiet=lowerCamelCase, echo=lowerCamelCase)) __lowerCAmelCase = ''' '''.join(lowerCamelCase) if result.returncode > 0: __lowerCAmelCase = '''\n'''.join(result.stderr) raise RuntimeError( F"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" F"""The combined stderr from workers follows:\n{stderr}""") return result class a__ ( __A ): """simple docstring""" pass def __magic_name__( lowerCamelCase, lowerCamelCase=False): try: __lowerCAmelCase = subprocess.check_output(lowerCamelCase, stderr=subprocess.STDOUT) if return_stdout: if hasattr(lowerCamelCase, '''decode'''): __lowerCAmelCase = output.decode('''utf-8''') return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F"""Command `{" ".join(lowerCamelCase)}` failed with the following error:\n\n{e.output.decode()}""") from e
368
'''simple docstring''' # Imports import numpy as np class a__ : """simple docstring""" def __init__(self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None ): self.set_matricies(red=__lowercase , green=__lowercase , blue=__lowercase , red_edge=__lowercase , nir=__lowercase ) def _snake_case (self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None ): if red is not None: __lowerCAmelCase = red if green is not None: __lowerCAmelCase = green if blue is not None: __lowerCAmelCase = blue if red_edge is not None: __lowerCAmelCase = red_edge if nir is not None: __lowerCAmelCase = nir return True def _snake_case (self , __lowercase="" , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None ): self.set_matricies(red=__lowercase , green=__lowercase , blue=__lowercase , red_edge=__lowercase , nir=__lowercase ) __lowerCAmelCase = { '''ARVI2''': self.arvaa, '''CCCI''': self.ccci, '''CVI''': self.cvi, '''GLI''': self.gli, '''NDVI''': self.ndvi, '''BNDVI''': self.bndvi, '''redEdgeNDVI''': self.red_edge_ndvi, '''GNDVI''': self.gndvi, '''GBNDVI''': self.gbndvi, '''GRNDVI''': self.grndvi, '''RBNDVI''': self.rbndvi, '''PNDVI''': self.pndvi, '''ATSAVI''': self.atsavi, '''BWDRVI''': self.bwdrvi, '''CIgreen''': self.ci_green, '''CIrededge''': self.ci_rededge, '''CI''': self.ci, '''CTVI''': self.ctvi, '''GDVI''': self.gdvi, '''EVI''': self.evi, '''GEMI''': self.gemi, '''GOSAVI''': self.gosavi, '''GSAVI''': self.gsavi, '''Hue''': self.hue, '''IVI''': self.ivi, '''IPVI''': self.ipvi, '''I''': self.i, '''RVI''': self.rvi, '''MRVI''': self.mrvi, '''MSAVI''': self.m_savi, '''NormG''': self.norm_g, '''NormNIR''': self.norm_nir, '''NormR''': self.norm_r, '''NGRDI''': self.ngrdi, '''RI''': self.ri, '''S''': self.s, '''IF''': self._if, '''DVI''': self.dvi, '''TVI''': self.tvi, '''NDRE''': self.ndre, } try: return funcs[index]() except KeyError: print('''Index not in the list!''' ) return False def _snake_case (self ): return -0.1_8 + (1.1_7 * ((self.nir - self.red) / (self.nir + self.red))) def _snake_case (self ): return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def _snake_case (self ): return self.nir * (self.red / (self.green**2)) def _snake_case (self ): return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def _snake_case (self ): return (self.nir - self.red) / (self.nir + self.red) def _snake_case (self ): return (self.nir - self.blue) / (self.nir + self.blue) def _snake_case (self ): return (self.redEdge - self.red) / (self.redEdge + self.red) def _snake_case (self ): return (self.nir - self.green) / (self.nir + self.green) def _snake_case (self ): return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def _snake_case (self ): return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def _snake_case (self ): return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def _snake_case (self ): return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def _snake_case (self , __lowercase=0.0_8 , __lowercase=1.2_2 , __lowercase=0.0_3 ): return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def _snake_case (self ): return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def _snake_case (self ): return (self.nir / self.green) - 1 def _snake_case (self ): return (self.nir / self.redEdge) - 1 def _snake_case (self ): return (self.red - self.blue) / self.red def _snake_case (self ): __lowerCAmelCase = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def _snake_case (self ): return self.nir - self.green def _snake_case (self ): return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def _snake_case (self ): __lowerCAmelCase = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.2_5 * n) - (self.red - 0.1_2_5) / (1 - self.red) def _snake_case (self , __lowercase=0.1_6 ): return (self.nir - self.green) / (self.nir + self.green + y) def _snake_case (self , __lowercase=0.5 ): return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def _snake_case (self ): return np.arctan( ((2 * self.red - self.green - self.blue) / 3_0.5) * (self.green - self.blue) ) def _snake_case (self , __lowercase=None , __lowercase=None ): return (self.nir - b) / (a * self.red) def _snake_case (self ): return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def _snake_case (self ): return (self.red + self.green + self.blue) / 3_0.5 def _snake_case (self ): return self.nir / self.red def _snake_case (self ): return (self.rvi() - 1) / (self.rvi() + 1) def _snake_case (self ): return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def _snake_case (self ): return self.green / (self.nir + self.red + self.green) def _snake_case (self ): return self.nir / (self.nir + self.red + self.green) def _snake_case (self ): return self.red / (self.nir + self.red + self.green) def _snake_case (self ): return (self.green - self.red) / (self.green + self.red) def _snake_case (self ): return (self.red - self.green) / (self.red + self.green) def _snake_case (self ): __lowerCAmelCase = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) __lowerCAmelCase = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def _snake_case (self ): return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def _snake_case (self ): return self.nir / self.red def _snake_case (self ): return (self.ndvi() + 0.5) ** (1 / 2) def _snake_case (self ): return (self.nir - self.redEdge) / (self.nir + self.redEdge)
9
0
'''simple docstring''' # XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path SCREAMING_SNAKE_CASE_: Any =Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) SCREAMING_SNAKE_CASE_: str ={'base': 'patrickvonplaten/wav2vec2_tiny_random', 'robust': 'patrickvonplaten/wav2vec2_tiny_random_robust'} SCREAMING_SNAKE_CASE_: Union[str, Any] ='zero2' SCREAMING_SNAKE_CASE_: Dict ='zero3' SCREAMING_SNAKE_CASE_: Optional[Any] =[ZEROa, ZEROa] def lowerCAmelCase_ ( snake_case_ : Optional[int] , snake_case_ : Optional[Any] , snake_case_ : Optional[int] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = parameterized.to_safe_name("_".join(str(snake_case_ ) for x in param.args ) ) return f"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test SCREAMING_SNAKE_CASE_: List[Any] =list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class __A ( UpperCamelCase__ ): @parameterized.expand(__a , name_func=__a ) def _lowercase (self : Union[str, Any] , __a : Dict , __a : Any ): self.run_and_check( stage=__a , model=__a , distributed=__a , fpaa=__a , ) @require_torch_multi_gpu @parameterized.expand(__a , name_func=__a ) def _lowercase (self : Tuple , __a : Dict , __a : List[Any] ): self.run_and_check( stage=__a , model=__a , distributed=__a , fpaa=__a , ) @parameterized.expand(__a , name_func=__a ) def _lowercase (self : Tuple , __a : Any , __a : str ): self.run_and_check( stage=__a , model=__a , distributed=__a , fpaa=__a , ) @require_torch_multi_gpu @parameterized.expand(__a , name_func=__a ) def _lowercase (self : int , __a : Union[str, Any] , __a : Any ): self.run_and_check( stage=__a , model=__a , distributed=__a , fpaa=__a , ) def _lowercase (self : str , __a : Optional[int] ): # XXX: run_asr is premature and doesn't save any results # so all we check for now is that the process didn't fail pass def _lowercase (self : List[str] , __a : str , __a : str , __a : int = 10 , __a : bool = True , __a : bool = True , __a : bool = True , ): UpperCAmelCase_ = models[model] UpperCAmelCase_ = self.run_trainer( stage=__a , model_name=__a , eval_steps=__a , num_train_epochs=1 , distributed=__a , fpaa=__a , ) self.do_checks(__a ) return output_dir def _lowercase (self : Any , __a : str , __a : str , __a : int = 10 , __a : int = 1 , __a : bool = True , __a : bool = True , ): UpperCAmelCase_ = self.get_auto_remove_tmp_dir("./xxx" , after=__a ) UpperCAmelCase_ = f""" --model_name_or_path {model_name} --dataset_name hf-internal-testing/librispeech_asr_dummy --dataset_config_name clean --train_split_name validation --validation_split_name validation --output_dir {output_dir} --num_train_epochs {str(__a )} --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --evaluation_strategy steps --learning_rate 5e-4 --warmup_steps 8 --orthography timit --preprocessing_num_workers 1 --group_by_length --freeze_feature_extractor --report_to none --save_steps 0 --eval_steps {eval_steps} --report_to none """.split() if fpaa: args.extend(["--fp16"] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files UpperCAmelCase_ = f"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() UpperCAmelCase_ = [f"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] UpperCAmelCase_ = self.get_launcher(__a ) UpperCAmelCase_ = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__a , env=self.get_env() ) return output_dir def _lowercase (self : Any , __a : List[str]=False ): # 1. explicitly set --num_nodes=1 just in case these tests end up run on a multi-node setup # - it won't be able to handle that # 2. for now testing with just 2 gpus max (since some quality tests may give different # results with mode gpus because we use very little data) UpperCAmelCase_ = min(2 , get_gpu_count() ) if distributed else 1 return f"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
1
import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class _lowerCamelCase ( a ): """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=768 ) -> List[str]: '''simple docstring''' super().__init__(UpperCAmelCase ) __snake_case : Optional[int] = proj_size __snake_case : str = CLIPVisionModel(UpperCAmelCase ) __snake_case : Tuple = PaintByExampleMapper(UpperCAmelCase ) __snake_case : Union[str, Any] = nn.LayerNorm(config.hidden_size ) __snake_case : Optional[Any] = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling __snake_case : Optional[int] = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase=False ) -> List[str]: '''simple docstring''' __snake_case : int = self.model(pixel_values=UpperCAmelCase ) __snake_case : Optional[int] = clip_output.pooler_output __snake_case : Any = self.mapper(latent_states[:, None] ) __snake_case : Any = self.final_layer_norm(UpperCAmelCase ) __snake_case : str = self.proj_out(UpperCAmelCase ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class _lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self , UpperCAmelCase ) -> List[Any]: '''simple docstring''' super().__init__() __snake_case : List[Any] = (config.num_hidden_layers + 1) // 5 __snake_case : Dict = config.hidden_size __snake_case : str = 1 __snake_case : List[Any] = nn.ModuleList( [ BasicTransformerBlock(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , activation_fn="gelu" , attention_bias=UpperCAmelCase ) for _ in range(UpperCAmelCase ) ] ) def UpperCAmelCase ( self , UpperCAmelCase ) -> str: '''simple docstring''' for block in self.blocks: __snake_case : int = block(UpperCAmelCase ) return hidden_states
326
0
from __future__ import annotations def _a ( _lowercase : list , _lowercase : int ): '''simple docstring''' if len(_lowercase ) <= 1 or n <= 1: return insert_next(_lowercase , n - 1 ) rec_insertion_sort(_lowercase , n - 1 ) def _a ( _lowercase : list , _lowercase : int ): '''simple docstring''' if index >= len(_lowercase ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order __UpperCAmelCase : int = ( collection[index], collection[index - 1], ) insert_next(_lowercase , index + 1 ) if __name__ == "__main__": __UpperCAmelCase :Any = input("Enter integers separated by spaces: ") __UpperCAmelCase :list[int] = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
368
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class a : """simple docstring""" def __init__( self : Union[str, Any] , snake_case : str , snake_case : Dict=13 , snake_case : Optional[Any]=7 , snake_case : Tuple=True , snake_case : Optional[int]=True , snake_case : str=True , snake_case : int=True , snake_case : List[str]=99 , snake_case : Any=32 , snake_case : List[str]=2 , snake_case : Tuple=4 , snake_case : Union[str, Any]=37 , snake_case : Dict="gelu" , snake_case : str=0.1 , snake_case : List[Any]=0.1 , snake_case : Any=512 , snake_case : Optional[Any]=16 , snake_case : Optional[int]=2 , snake_case : Union[str, Any]=0.02 , snake_case : List[Any]=3 , snake_case : str=4 , snake_case : int=None , snake_case : Union[str, Any]=1000 , ) -> Tuple: __UpperCAmelCase : int = parent __UpperCAmelCase : Optional[int] = batch_size __UpperCAmelCase : Dict = seq_length __UpperCAmelCase : List[Any] = is_training __UpperCAmelCase : Optional[Any] = use_input_mask __UpperCAmelCase : List[Any] = use_token_type_ids __UpperCAmelCase : str = use_labels __UpperCAmelCase : Any = vocab_size __UpperCAmelCase : List[str] = hidden_size __UpperCAmelCase : Dict = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : List[str] = intermediate_size __UpperCAmelCase : str = hidden_act __UpperCAmelCase : int = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : List[str] = max_position_embeddings __UpperCAmelCase : str = type_vocab_size __UpperCAmelCase : Dict = type_sequence_label_size __UpperCAmelCase : Any = initializer_range __UpperCAmelCase : Optional[int] = num_labels __UpperCAmelCase : Optional[Any] = num_choices __UpperCAmelCase : List[Any] = scope __UpperCAmelCase : str = range_bbox def lowerCamelCase__ ( self : Tuple ) -> Union[str, Any]: __UpperCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __UpperCAmelCase : Optional[int] = bbox[i, j, 3] __UpperCAmelCase : Any = bbox[i, j, 1] __UpperCAmelCase : List[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: __UpperCAmelCase : str = bbox[i, j, 2] __UpperCAmelCase : List[Any] = bbox[i, j, 0] __UpperCAmelCase : Dict = t __UpperCAmelCase : Any = tf.convert_to_tensor(snake_case ) __UpperCAmelCase : List[Any] = None if self.use_input_mask: __UpperCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : int = None if self.use_token_type_ids: __UpperCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : Optional[Any] = None __UpperCAmelCase : Optional[int] = None __UpperCAmelCase : str = None if self.use_labels: __UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) __UpperCAmelCase : Optional[int] = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self : List[str] , snake_case : int , snake_case : str , snake_case : Tuple , snake_case : List[str] , snake_case : Any , snake_case : Any , snake_case : List[Any] , snake_case : Any ) -> Optional[Any]: __UpperCAmelCase : Tuple = TFLayoutLMModel(config=snake_case ) __UpperCAmelCase : Optional[Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) __UpperCAmelCase : Tuple = model(snake_case , snake_case , token_type_ids=snake_case ) __UpperCAmelCase : List[Any] = model(snake_case , snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self : Optional[Any] , snake_case : Union[str, Any] , snake_case : List[str] , snake_case : Dict , snake_case : str , snake_case : Optional[int] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : str ) -> int: __UpperCAmelCase : Any = TFLayoutLMForMaskedLM(config=snake_case ) __UpperCAmelCase : List[Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self : Tuple , snake_case : Any , snake_case : Dict , snake_case : str , snake_case : Tuple , snake_case : str , snake_case : Optional[Any] , snake_case : str , snake_case : str ) -> Any: __UpperCAmelCase : List[str] = self.num_labels __UpperCAmelCase : Optional[int] = TFLayoutLMForSequenceClassification(config=snake_case ) __UpperCAmelCase : Any = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self : Dict , snake_case : List[str] , snake_case : Dict , snake_case : Union[str, Any] , snake_case : List[Any] , snake_case : Union[str, Any] , snake_case : Any , snake_case : Tuple , snake_case : List[str] ) -> List[str]: __UpperCAmelCase : Tuple = self.num_labels __UpperCAmelCase : Optional[int] = TFLayoutLMForTokenClassification(config=snake_case ) __UpperCAmelCase : Any = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self : int , snake_case : Dict , snake_case : int , snake_case : Union[str, Any] , snake_case : List[str] , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Dict , snake_case : Optional[int] ) -> Dict: __UpperCAmelCase : int = TFLayoutLMForQuestionAnswering(config=snake_case ) __UpperCAmelCase : Union[str, Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self : Dict ) -> List[str]: __UpperCAmelCase : str = self.prepare_config_and_inputs() ( ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ) : Tuple = config_and_inputs __UpperCAmelCase : Any = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_tf class a ( _a , _a , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Optional[int] = ( { "feature-extraction": TFLayoutLMModel, "fill-mask": TFLayoutLMForMaskedLM, "text-classification": TFLayoutLMForSequenceClassification, "token-classification": TFLayoutLMForTokenClassification, "zero-shot": TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : Any = True SCREAMING_SNAKE_CASE : List[str] = 1_0 def lowerCamelCase__ ( self : Optional[Any] ) -> Any: __UpperCAmelCase : Optional[int] = TFLayoutLMModelTester(self ) __UpperCAmelCase : Dict = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def lowerCamelCase__ ( self : Any ) -> Dict: self.config_tester.run_common_tests() def lowerCamelCase__ ( self : Dict ) -> List[str]: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def lowerCamelCase__ ( self : Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def lowerCamelCase__ ( self : List[Any] ) -> Any: __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def lowerCamelCase__ ( self : int ) -> List[Any]: __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) def lowerCamelCase__ ( self : Dict ) -> List[Any]: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) @slow def lowerCamelCase__ ( self : List[str] ) -> Union[str, Any]: for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase : int = TFLayoutLMModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip('''Onnx compliancy broke with TF 2.10''' ) def lowerCamelCase__ ( self : Dict ) -> Dict: pass def _a ( ): '''simple docstring''' __UpperCAmelCase : str = tf.convert_to_tensor([[101,1019,1014,1016,1037,12849,4747,1004,14246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,11300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,19274,2772,6205,27814,16147,16147,4343,2047,10283,10969,14389,1012,2338,102]] ) # noqa: E231 __UpperCAmelCase : Optional[Any] = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 __UpperCAmelCase : Optional[Any] = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 __UpperCAmelCase : str = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) __UpperCAmelCase : Optional[int] = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class a ( unittest.TestCase ): """simple docstring""" @slow def lowerCamelCase__ ( self : List[str] ) -> Optional[int]: __UpperCAmelCase : int = TFLayoutLMModel.from_pretrained('''microsoft/layoutlm-base-uncased''' ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : str = prepare_layoutlm_batch_inputs() # forward pass __UpperCAmelCase : Dict = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the sequence output on [0, :3, :3] __UpperCAmelCase : Union[str, Any] = tf.convert_to_tensor( [[0.1_785, -0.1_947, -0.0_425], [-0.3_254, -0.2_807, 0.2_553], [-0.5_391, -0.3_322, 0.3_364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case , atol=1E-3 ) ) # test the pooled output on [1, :3] __UpperCAmelCase : str = tf.convert_to_tensor([-0.6_580, -0.0_214, 0.8_552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , snake_case , atol=1E-3 ) ) @slow def lowerCamelCase__ ( self : Optional[int] ) -> int: # initialize model with randomly initialized sequence classification head __UpperCAmelCase : str = TFLayoutLMForSequenceClassification.from_pretrained('''microsoft/layoutlm-base-uncased''' , num_labels=2 ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : int = prepare_layoutlm_batch_inputs() # forward pass __UpperCAmelCase : Tuple = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar __UpperCAmelCase : str = outputs.loss __UpperCAmelCase : Optional[Any] = (2,) self.assertEqual(loss.shape , snake_case ) # test the shape of the logits __UpperCAmelCase : List[str] = outputs.logits __UpperCAmelCase : List[Any] = (2, 2) self.assertEqual(logits.shape , snake_case ) @slow def lowerCamelCase__ ( self : List[Any] ) -> str: # initialize model with randomly initialized token classification head __UpperCAmelCase : Union[str, Any] = TFLayoutLMForTokenClassification.from_pretrained('''microsoft/layoutlm-base-uncased''' , num_labels=13 ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = prepare_layoutlm_batch_inputs() # forward pass __UpperCAmelCase : Tuple = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) # test the shape of the logits __UpperCAmelCase : List[Any] = outputs.logits __UpperCAmelCase : Optional[int] = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , snake_case ) @slow def lowerCamelCase__ ( self : Optional[Any] ) -> Union[str, Any]: # initialize model with randomly initialized token classification head __UpperCAmelCase : Dict = TFLayoutLMForQuestionAnswering.from_pretrained('''microsoft/layoutlm-base-uncased''' ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = prepare_layoutlm_batch_inputs() # forward pass __UpperCAmelCase : Optional[Any] = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the shape of the logits __UpperCAmelCase : Union[str, Any] = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , snake_case ) self.assertEqual(outputs.end_logits.shape , snake_case )
240
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class UpperCAmelCase_ ( unittest.TestCase): def __init__( self : Dict , __UpperCamelCase : Optional[int] , __UpperCamelCase : List[str]=7 , __UpperCamelCase : Any=3 , __UpperCamelCase : Any=30 , __UpperCamelCase : Tuple=400 , __UpperCamelCase : int=True , __UpperCamelCase : Tuple=None , __UpperCamelCase : Optional[Any]=0.9 , __UpperCamelCase : int=None , __UpperCamelCase : Optional[Any]=True , __UpperCamelCase : Tuple=[0.5, 0.5, 0.5] , __UpperCamelCase : str=[0.5, 0.5, 0.5] , ) -> Tuple: _UpperCamelCase = size if size is not None else {'''shortest_edge''': 30} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 30, '''width''': 30} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize_and_center_crop _UpperCamelCase = size _UpperCamelCase = crop_pct _UpperCamelCase = crop_size _UpperCamelCase = do_normalize _UpperCamelCase = image_mean _UpperCamelCase = image_std def _UpperCamelCase ( self : Any ) -> Optional[int]: return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class UpperCAmelCase_ ( _lowercase , unittest.TestCase): snake_case__ = PoolFormerImageProcessor if is_vision_available() else None def _UpperCamelCase ( self : int ) -> str: _UpperCamelCase = PoolFormerImageProcessingTester(self ) @property def _UpperCamelCase ( self : List[Any] ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self : List[Any] ) -> Dict: _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCamelCase , '''do_resize_and_center_crop''' ) ) self.assertTrue(hasattr(__UpperCamelCase , '''size''' ) ) self.assertTrue(hasattr(__UpperCamelCase , '''crop_pct''' ) ) self.assertTrue(hasattr(__UpperCamelCase , '''do_normalize''' ) ) self.assertTrue(hasattr(__UpperCamelCase , '''image_mean''' ) ) self.assertTrue(hasattr(__UpperCamelCase , '''image_std''' ) ) def _UpperCamelCase ( self : Any ) -> int: _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 30} ) self.assertEqual(image_processor.crop_size , {'''height''': 30, '''width''': 30} ) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def _UpperCamelCase ( self : Any ) -> List[str]: pass def _UpperCamelCase ( self : Optional[Any] ) -> Optional[Any]: # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCamelCase ) for image in image_inputs: self.assertIsInstance(__UpperCamelCase , Image.Image ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def _UpperCamelCase ( self : Optional[int] ) -> str: # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCamelCase , numpify=__UpperCamelCase ) for image in image_inputs: self.assertIsInstance(__UpperCamelCase , np.ndarray ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def _UpperCamelCase ( self : str ) -> List[Any]: # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCamelCase , torchify=__UpperCamelCase ) for image in image_inputs: self.assertIsInstance(__UpperCamelCase , torch.Tensor ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
256
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"""vocab_file""": """spiece.model"""} UpperCAmelCase = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } UpperCAmelCase = {"""bert_for_seq_generation""": 512} class UpperCAmelCase_ ( _lowercase): snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = [] snake_case__ = ['''input_ids''', '''attention_mask'''] def __init__( self : Any , __UpperCamelCase : int , __UpperCamelCase : Optional[int]="<s>" , __UpperCamelCase : Optional[Any]="</s>" , __UpperCamelCase : Optional[Any]="<unk>" , __UpperCamelCase : Tuple="<pad>" , __UpperCamelCase : int="<::::>" , __UpperCamelCase : Optional[Dict[str, Any]] = None , **__UpperCamelCase : Any , ) -> None: _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , pad_token=__UpperCamelCase , sep_token=__UpperCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCamelCase , ) _UpperCamelCase = vocab_file _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCamelCase ) @property def _UpperCamelCase ( self : Optional[int] ) -> Tuple: return self.sp_model.get_piece_size() def _UpperCamelCase ( self : int ) -> Optional[int]: _UpperCamelCase = {self.convert_ids_to_tokens(__UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ) -> Union[str, Any]: _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : str , __UpperCamelCase : Any ) -> Tuple: _UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : Union[str, Any] , __UpperCamelCase : str ) -> List[str]: return self.sp_model.encode(__UpperCamelCase , out_type=__UpperCamelCase ) def _UpperCamelCase ( self : Tuple , __UpperCamelCase : Any ) -> Optional[int]: return self.sp_model.piece_to_id(__UpperCamelCase ) def _UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : Optional[int] ) -> Optional[Any]: _UpperCamelCase = self.sp_model.IdToPiece(__UpperCamelCase ) return token def _UpperCamelCase ( self : str , __UpperCamelCase : Dict ) -> Optional[Any]: _UpperCamelCase = [] _UpperCamelCase = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__UpperCamelCase ) + token _UpperCamelCase = [] else: current_sub_tokens.append(__UpperCamelCase ) out_string += self.sp_model.decode(__UpperCamelCase ) return out_string.strip() def _UpperCamelCase ( self : Tuple , __UpperCamelCase : str , __UpperCamelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__UpperCamelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCamelCase = os.path.join( __UpperCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCamelCase , '''wb''' ) as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCamelCase ) return (out_vocab_file,)
256
1
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase: List[str] = (7_2_0, 1_2_8_0) # Height, Width lowerCAmelCase: str = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase: List[str] = 1 / 1_0_0 lowerCAmelCase: Dict = '' lowerCAmelCase: Any = '' lowerCAmelCase: str = '' lowerCAmelCase: Union[str, Any] = 2_5_0 def lowerCamelCase__ ( ): a : Any = get_dataset(_A , _A ) for index in range(_A ): a : List[str] = random.sample(range(len(_A ) ) , 4 ) a : List[Any] = update_image_and_anno( _A , _A , _A , _A , _A , filter_scale=_A , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' a : str = random_chars(32 ) a : Tuple = path.split(os.sep )[-1].rsplit('.' , 1 )[0] a : str = f"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(f"""{file_root}.jpg""" , _A , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) a : List[str] = [] for anno in new_annos: a : Optional[Any] = anno[3] - anno[1] a : Dict = anno[4] - anno[2] a : List[str] = anno[1] + width / 2 a : Union[str, Any] = anno[2] + height / 2 a : int = f"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(_A ) with open(f"""{file_root}.txt""" , 'w' ) as outfile: outfile.write('\n'.join(line for line in annos_list ) ) def lowerCamelCase__ ( _A , _A ): a : Optional[int] = [] a : Dict = [] for label_file in glob.glob(os.path.join(_A , '*.txt' ) ): a : int = label_file.split(os.sep )[-1].rsplit('.' , 1 )[0] with open(_A ) as in_file: a : Tuple = in_file.readlines() a : List[str] = os.path.join(_A , f"""{label_name}.jpg""" ) a : Optional[int] = [] for obj_list in obj_lists: a : Tuple = obj_list.rstrip('\n' ).split(' ' ) a : List[str] = float(obj[1] ) - float(obj[3] ) / 2 a : List[str] = float(obj[2] ) - float(obj[4] ) / 2 a : Any = float(obj[1] ) + float(obj[3] ) / 2 a : str = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(_A ) labels.append(_A ) return img_paths, labels def lowerCamelCase__ ( _A , _A , _A , _A , _A , _A = 0.0 , ): a : Any = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) a : Union[str, Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) a : Union[str, Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) a : Any = int(scale_x * output_size[1] ) a : List[Any] = int(scale_y * output_size[0] ) a : Dict = [] a : Any = [] for i, index in enumerate(_A ): a : List[Any] = all_img_list[index] path_list.append(_A ) a : str = all_annos[index] a : str = cva.imread(_A ) if i == 0: # top-left a : Any = cva.resize(_A , (divid_point_x, divid_point_y) ) a : Optional[Any] = img for bbox in img_annos: a : Any = bbox[1] * scale_x a : Union[str, Any] = bbox[2] * scale_y a : int = bbox[3] * scale_x a : Dict = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right a : Union[str, Any] = cva.resize(_A , (output_size[1] - divid_point_x, divid_point_y) ) a : Optional[Any] = img for bbox in img_annos: a : List[Any] = scale_x + bbox[1] * (1 - scale_x) a : List[Any] = bbox[2] * scale_y a : Optional[int] = scale_x + bbox[3] * (1 - scale_x) a : Optional[int] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left a : int = cva.resize(_A , (divid_point_x, output_size[0] - divid_point_y) ) a : Optional[int] = img for bbox in img_annos: a : int = bbox[1] * scale_x a : Optional[int] = scale_y + bbox[2] * (1 - scale_y) a : int = bbox[3] * scale_x a : Dict = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right a : List[str] = cva.resize( _A , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) a : Any = img for bbox in img_annos: a : Optional[int] = scale_x + bbox[1] * (1 - scale_x) a : str = scale_y + bbox[2] * (1 - scale_y) a : Dict = scale_x + bbox[3] * (1 - scale_x) a : int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: a : Optional[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def lowerCamelCase__ ( _A ): assert number_char > 1, "The number of character should greater than 1" a : List[Any] = ascii_lowercase + digits return "".join(random.choice(_A ) for _ in range(_A ) ) if __name__ == "__main__": main() print('DONE ✅')
360
'''simple docstring''' import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class a__( lowerCamelCase__ , unittest.TestCase ): lowercase__ = CTRLTokenizer lowercase__ = False lowercase__ = False def lowercase_ ( self : Dict ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt a : Tuple = ['adapt', 're@@', 'a@@', 'apt', 'c@@', 't', '<unk>'] a : Union[str, Any] = dict(zip(__snake_case , range(len(__snake_case ) ) ) ) a : Union[str, Any] = ['#version: 0.2', 'a p', 'ap t</w>', 'r e', 'a d', 'ad apt</w>', ''] a : Optional[Any] = {'unk_token': '<unk>'} a : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) a : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__snake_case ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__snake_case ) ) def lowercase_ ( self : int , **__snake_case : str ): kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **__snake_case ) def lowercase_ ( self : Optional[int] , __snake_case : Any ): a : int = 'adapt react readapt apt' a : Any = 'adapt react readapt apt' return input_text, output_text def lowercase_ ( self : Dict ): a : Dict = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) a : List[str] = 'adapt react readapt apt' a : Dict = 'adapt re@@ a@@ c@@ t re@@ adapt apt'.split() a : Any = tokenizer.tokenize(__snake_case ) self.assertListEqual(__snake_case , __snake_case ) a : Dict = tokens + [tokenizer.unk_token] a : Optional[Any] = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(__snake_case ) , __snake_case )
96
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, require_torch_gpu, require_torch_or_tf, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf class _snake_case ( unittest.TestCase ): lowerCAmelCase_ : Optional[Any] = MODEL_FOR_CAUSAL_LM_MAPPING lowerCAmelCase_ : Optional[Any] = TF_MODEL_FOR_CAUSAL_LM_MAPPING @require_torch def lowerCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ = pipeline(task="text-generation" , model="sshleifer/tiny-ctrl" , framework="pt" ) # Using `do_sample=False` to force deterministic output snake_case_ = text_generator("This is a test" , do_sample=a__ ) self.assertEqual( a__ , [ { "generated_text": ( "This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope." " oscope. FiliFili@@" ) } ] , ) snake_case_ = text_generator(["This is a test", "This is a second test"] ) self.assertEqual( a__ , [ [ { "generated_text": ( "This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope." " oscope. FiliFili@@" ) } ], [ { "generated_text": ( "This is a second test ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy" " oscope. oscope. FiliFili@@" ) } ], ] , ) snake_case_ = text_generator("This is a test" , do_sample=a__ , num_return_sequences=2 , return_tensors=a__ ) self.assertEqual( a__ , [ {"generated_token_ids": ANY(a__ )}, {"generated_token_ids": ANY(a__ )}, ] , ) snake_case_ = text_generator.model.config.eos_token_id snake_case_ = "<pad>" snake_case_ = text_generator( ["This is a test", "This is a second test"] , do_sample=a__ , num_return_sequences=2 , batch_size=2 , return_tensors=a__ , ) self.assertEqual( a__ , [ [ {"generated_token_ids": ANY(a__ )}, {"generated_token_ids": ANY(a__ )}, ], [ {"generated_token_ids": ANY(a__ )}, {"generated_token_ids": ANY(a__ )}, ], ] , ) @require_tf def lowerCAmelCase__ ( self ) -> Tuple: '''simple docstring''' snake_case_ = pipeline(task="text-generation" , model="sshleifer/tiny-ctrl" , framework="tf" ) # Using `do_sample=False` to force deterministic output snake_case_ = text_generator("This is a test" , do_sample=a__ ) self.assertEqual( a__ , [ { "generated_text": ( "This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵" " please," ) } ] , ) snake_case_ = text_generator(["This is a test", "This is a second test"] , do_sample=a__ ) self.assertEqual( a__ , [ [ { "generated_text": ( "This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵" " please," ) } ], [ { "generated_text": ( "This is a second test Chieftain Chieftain prefecture prefecture prefecture Cannes Cannes" " Cannes 閲閲Cannes Cannes Cannes 攵 please," ) } ], ] , ) def lowerCAmelCase__ ( self , a__ , a__ , a__ ) -> str: '''simple docstring''' snake_case_ = TextGenerationPipeline(model=a__ , tokenizer=a__ ) return text_generator, ["This is a test", "Another test"] def lowerCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ = "Hello I believe in" snake_case_ = pipeline("text-generation" , model="hf-internal-testing/tiny-random-gpt2" ) snake_case_ = text_generator(a__ ) self.assertEqual( a__ , [{"generated_text": "Hello I believe in fe fe fe fe fe fe fe fe fe fe fe fe"}] , ) snake_case_ = text_generator(a__ , stop_sequence=" fe" ) self.assertEqual(a__ , [{"generated_text": "Hello I believe in fe"}] ) def lowerCAmelCase__ ( self , a__ , a__ ) -> Tuple: '''simple docstring''' snake_case_ = text_generator.model snake_case_ = text_generator.tokenizer snake_case_ = text_generator("This is a test" ) self.assertEqual(a__ , [{"generated_text": ANY(a__ )}] ) self.assertTrue(outputs[0]["generated_text"].startswith("This is a test" ) ) snake_case_ = text_generator("This is a test" , return_full_text=a__ ) self.assertEqual(a__ , [{"generated_text": ANY(a__ )}] ) self.assertNotIn("This is a test" , outputs[0]["generated_text"] ) snake_case_ = pipeline(task="text-generation" , model=a__ , tokenizer=a__ , return_full_text=a__ ) snake_case_ = text_generator("This is a test" ) self.assertEqual(a__ , [{"generated_text": ANY(a__ )}] ) self.assertNotIn("This is a test" , outputs[0]["generated_text"] ) snake_case_ = text_generator("This is a test" , return_full_text=a__ ) self.assertEqual(a__ , [{"generated_text": ANY(a__ )}] ) self.assertTrue(outputs[0]["generated_text"].startswith("This is a test" ) ) snake_case_ = text_generator(["This is great !", "Something else"] , num_return_sequences=2 , do_sample=a__ ) self.assertEqual( a__ , [ [{"generated_text": ANY(a__ )}, {"generated_text": ANY(a__ )}], [{"generated_text": ANY(a__ )}, {"generated_text": ANY(a__ )}], ] , ) if text_generator.tokenizer.pad_token is not None: snake_case_ = text_generator( ["This is great !", "Something else"] , num_return_sequences=2 , batch_size=2 , do_sample=a__ ) self.assertEqual( a__ , [ [{"generated_text": ANY(a__ )}, {"generated_text": ANY(a__ )}], [{"generated_text": ANY(a__ )}, {"generated_text": ANY(a__ )}], ] , ) with self.assertRaises(a__ ): snake_case_ = text_generator("test" , return_full_text=a__ , return_text=a__ ) with self.assertRaises(a__ ): snake_case_ = text_generator("test" , return_full_text=a__ , return_tensors=a__ ) with self.assertRaises(a__ ): snake_case_ = text_generator("test" , return_text=a__ , return_tensors=a__ ) # Empty prompt is slighly special # it requires BOS token to exist. # Special case for Pegasus which will always append EOS so will # work even without BOS. if ( text_generator.tokenizer.bos_token_id is not None or "Pegasus" in tokenizer.__class__.__name__ or "Git" in model.__class__.__name__ ): snake_case_ = text_generator("" ) self.assertEqual(a__ , [{"generated_text": ANY(a__ )}] ) else: with self.assertRaises((ValueError, AssertionError) ): snake_case_ = text_generator("" ) if text_generator.framework == "tf": # TF generation does not support max_new_tokens, and it's impossible # to control long generation with only max_length without # fancy calculation, dismissing tests for now. return # We don't care about infinite range models. # They already work. # Skip this test for XGLM, since it uses sinusoidal positional embeddings which are resized on-the-fly. snake_case_ = ["RwkvForCausalLM", "XGLMForCausalLM", "GPTNeoXForCausalLM"] if ( tokenizer.model_max_length < 10_000 and text_generator.model.__class__.__name__ not in EXTRA_MODELS_CAN_HANDLE_LONG_INPUTS ): # Handling of large generations with self.assertRaises((RuntimeError, IndexError, ValueError, AssertionError) ): text_generator("This is a test" * 500 , max_new_tokens=20 ) snake_case_ = text_generator("This is a test" * 500 , handle_long_generation="hole" , max_new_tokens=20 ) # Hole strategy cannot work with self.assertRaises(a__ ): text_generator( "This is a test" * 500 , handle_long_generation="hole" , max_new_tokens=tokenizer.model_max_length + 10 , ) @require_torch @require_accelerate @require_torch_gpu def lowerCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' import torch # Classic `model_kwargs` snake_case_ = pipeline( model="hf-internal-testing/tiny-random-bloom" , model_kwargs={"device_map": "auto", "torch_dtype": torch.bfloataa} , ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) snake_case_ = pipe("This is a test" ) self.assertEqual( a__ , [ { "generated_text": ( "This is a test test test test test test test test test test test test test test test test" " test" ) } ] , ) # Upgraded those two to real pipeline arguments (they just get sent for the model as they're unlikely to mean anything else.) snake_case_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device_map="auto" , torch_dtype=torch.bfloataa ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) snake_case_ = pipe("This is a test" ) self.assertEqual( a__ , [ { "generated_text": ( "This is a test test test test test test test test test test test test test test test test" " test" ) } ] , ) # torch_dtype will be automatically set to float32 if not provided - check: https://github.com/huggingface/transformers/pull/20602 snake_case_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device_map="auto" ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.floataa ) snake_case_ = pipe("This is a test" ) self.assertEqual( a__ , [ { "generated_text": ( "This is a test test test test test test test test test test test test test test test test" " test" ) } ] , ) @require_torch @require_torch_gpu def lowerCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' import torch snake_case_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device=0 , torch_dtype=torch.floataa ) pipe("This is a test" ) @require_torch @require_accelerate @require_torch_gpu def lowerCAmelCase__ ( self ) -> Dict: '''simple docstring''' import torch snake_case_ = pipeline(model="hf-internal-testing/tiny-random-bloom" , device_map="auto" , torch_dtype=torch.floataa ) pipe("This is a test" , do_sample=a__ , top_p=0.5 ) def lowerCAmelCase__ ( self ) -> Tuple: '''simple docstring''' snake_case_ = "Hello world" snake_case_ = pipeline("text-generation" , model="hf-internal-testing/tiny-random-gpt2" ) if text_generator.model.framework == "tf": snake_case_ = logging.get_logger("transformers.generation.tf_utils" ) else: snake_case_ = logging.get_logger("transformers.generation.utils" ) snake_case_ = "Both `max_new_tokens`" # The beggining of the message to be checked in this test # Both are set by the user -> log warning with CaptureLogger(a__ ) as cl: snake_case_ = text_generator(a__ , max_length=10 , max_new_tokens=1 ) self.assertIn(a__ , cl.out ) # The user only sets one -> no warning with CaptureLogger(a__ ) as cl: snake_case_ = text_generator(a__ , max_new_tokens=1 ) self.assertNotIn(a__ , cl.out ) with CaptureLogger(a__ ) as cl: snake_case_ = text_generator(a__ , max_length=10 ) self.assertNotIn(a__ , cl.out )
85
'''simple docstring''' import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP _SCREAMING_SNAKE_CASE : Any = False try: _SCREAMING_SNAKE_CASE : Optional[Any] = _is_package_available("google.colab") except ModuleNotFoundError: pass @input.register class _snake_case : def __init__( self , a__ = None , a__ = [] ) -> List[str]: '''simple docstring''' snake_case_ = 0 snake_case_ = choices snake_case_ = prompt if sys.platform == "win32": snake_case_ = "*" else: snake_case_ = "➔ " def lowerCAmelCase__ ( self , a__ , a__ = "" ) -> int: '''simple docstring''' if sys.platform != "win32": writeColor(self.choices[index] , 32 , a__ ) else: forceWrite(self.choices[index] , a__ ) def lowerCAmelCase__ ( self , a__ ) -> Tuple: '''simple docstring''' if index == self.position: forceWrite(F' {self.arrow_char} ' ) self.write_choice(a__ ) else: forceWrite(F' {self.choices[index]}' ) reset_cursor() def lowerCAmelCase__ ( self , a__ , a__ = 1 ) -> List[str]: '''simple docstring''' snake_case_ = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(a__ ) move_cursor(a__ , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP["up"] ) def lowerCAmelCase__ ( self ) -> Dict: '''simple docstring''' self.move_direction(Direction.UP ) @input.mark(KEYMAP["down"] ) def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' self.move_direction(Direction.DOWN ) @input.mark(KEYMAP["newline"] ) def lowerCAmelCase__ ( self ) -> str: '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) return self.position @input.mark(KEYMAP["interrupt"] ) def lowerCAmelCase__ ( self ) -> Tuple: '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(a__ )] for number in range(10 )] ) def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ = int(chr(self.current_selection ) ) snake_case_ = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , a__ ) else: return else: return def lowerCAmelCase__ ( self , a__ = 0 ) -> List[str]: '''simple docstring''' if self.prompt: linebreak() forceWrite(self.prompt , "\n" ) if in_colab: forceWrite("Please input a choice index (starting from 0), and press enter" , "\n" ) else: forceWrite("Please select a choice using the arrow or number keys, and selecting with enter" , "\n" ) snake_case_ = default_choice for i in range(len(self.choices ) ): self.print_choice(a__ ) forceWrite("\n" ) move_cursor(len(self.choices ) - self.position , "UP" ) with cursor.hide(): while True: if in_colab: try: snake_case_ = int(builtins.input() ) except ValueError: snake_case_ = default_choice else: snake_case_ = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , "UP" ) clear_line() self.write_choice(a__ , "\n" ) return choice
85
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { 'google/vit-base-patch16-224': 'https://huggingface.co/vit-base-patch16-224/resolve/main/config.json', # See all ViT models at https://huggingface.co/models?filter=vit } class __A ( A_ ): '''simple docstring''' lowerCAmelCase : Any = "vit" def __init__( self : List[str] ,_snake_case : str=768 ,_snake_case : Optional[Any]=12 ,_snake_case : Dict=12 ,_snake_case : Dict=3_072 ,_snake_case : int="gelu" ,_snake_case : Optional[int]=0.0 ,_snake_case : List[str]=0.0 ,_snake_case : str=0.02 ,_snake_case : Optional[Any]=1e-12 ,_snake_case : Optional[Any]=224 ,_snake_case : Optional[Any]=16 ,_snake_case : List[Any]=3 ,_snake_case : Optional[int]=True ,_snake_case : Any=16 ,**_snake_case : Optional[int] ,) -> Any: """simple docstring""" super().__init__(**_snake_case ) lowercase__ : Optional[Any] = hidden_size lowercase__ : Dict = num_hidden_layers lowercase__ : Dict = num_attention_heads lowercase__ : Tuple = intermediate_size lowercase__ : Dict = hidden_act lowercase__ : str = hidden_dropout_prob lowercase__ : Optional[int] = attention_probs_dropout_prob lowercase__ : Optional[Any] = initializer_range lowercase__ : Tuple = layer_norm_eps lowercase__ : Tuple = image_size lowercase__ : str = patch_size lowercase__ : str = num_channels lowercase__ : Optional[int] = qkv_bias lowercase__ : Tuple = encoder_stride class __A ( A_ ): '''simple docstring''' lowerCAmelCase : Dict = version.parse("1.11" ) @property def UpperCAmelCase ( self : Any ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def UpperCAmelCase ( self : Union[str, Any] ) -> float: """simple docstring""" return 1e-4
357
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig lowerCAmelCase_ = [ 'openmmlab/upernet-convnext-tiny', # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring lowerCAmelCase_ = 'UperNetConfig' class __A ( nn.Module ): '''simple docstring''' def __init__( self : List[str] ,_snake_case : int ,_snake_case : int ,_snake_case : Union[int, Tuple[int, int]] ,_snake_case : Union[int, Tuple[int, int], str] = 0 ,_snake_case : bool = False ,_snake_case : Union[int, Tuple[int, int]] = 1 ,) -> None: """simple docstring""" super().__init__() lowercase__ : Optional[int] = nn.Convad( in_channels=_snake_case ,out_channels=_snake_case ,kernel_size=_snake_case ,padding=_snake_case ,bias=_snake_case ,dilation=_snake_case ,) lowercase__ : Tuple = nn.BatchNormad(_snake_case ) lowercase__ : List[str] = nn.ReLU() def UpperCAmelCase ( self : str ,_snake_case : torch.Tensor ) -> torch.Tensor: """simple docstring""" lowercase__ : Union[str, Any] = self.conv(_snake_case ) lowercase__ : List[str] = self.batch_norm(_snake_case ) lowercase__ : Tuple = self.activation(_snake_case ) return output class __A ( nn.Module ): '''simple docstring''' def __init__( self : Union[str, Any] ,_snake_case : int ,_snake_case : int ,_snake_case : int ) -> None: """simple docstring""" super().__init__() lowercase__ : List[Any] = [ nn.AdaptiveAvgPoolad(_snake_case ), UperNetConvModule(_snake_case ,_snake_case ,kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(_snake_case ) ,_snake_case ) def UpperCAmelCase ( self : Dict ,_snake_case : torch.Tensor ) -> torch.Tensor: """simple docstring""" lowercase__ : Any = input for layer in self.layers: lowercase__ : int = layer(_snake_case ) return hidden_state class __A ( nn.Module ): '''simple docstring''' def __init__( self : List[str] ,_snake_case : Tuple[int, ...] ,_snake_case : int ,_snake_case : int ,_snake_case : bool ) -> None: """simple docstring""" super().__init__() lowercase__ : int = pool_scales lowercase__ : Dict = align_corners lowercase__ : Optional[Any] = in_channels lowercase__ : Optional[Any] = channels lowercase__ : int = [] for i, pool_scale in enumerate(_snake_case ): lowercase__ : Optional[Any] = UperNetPyramidPoolingBlock(pool_scale=_snake_case ,in_channels=_snake_case ,channels=_snake_case ) self.blocks.append(_snake_case ) self.add_module(str(_snake_case ) ,_snake_case ) def UpperCAmelCase ( self : Any ,_snake_case : torch.Tensor ) -> List[torch.Tensor]: """simple docstring""" lowercase__ : int = [] for ppm in self.blocks: lowercase__ : Any = ppm(_snake_case ) lowercase__ : int = nn.functional.interpolate( _snake_case ,size=x.size()[2:] ,mode='''bilinear''' ,align_corners=self.align_corners ) ppm_outs.append(_snake_case ) return ppm_outs class __A ( nn.Module ): '''simple docstring''' def __init__( self : Union[str, Any] ,_snake_case : List[str] ,_snake_case : Union[str, Any] ) -> str: """simple docstring""" super().__init__() lowercase__ : str = config lowercase__ : Optional[Any] = config.pool_scales # e.g. (1, 2, 3, 6) lowercase__ : Optional[Any] = in_channels lowercase__ : Any = config.hidden_size lowercase__ : Optional[Any] = False lowercase__ : Optional[int] = nn.Convad(self.channels ,config.num_labels ,kernel_size=1 ) # PSP Module lowercase__ : Dict = UperNetPyramidPoolingModule( self.pool_scales ,self.in_channels[-1] ,self.channels ,align_corners=self.align_corners ,) lowercase__ : str = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels ,self.channels ,kernel_size=3 ,padding=1 ,) # FPN Module lowercase__ : Any = nn.ModuleList() lowercase__ : Union[str, Any] = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer lowercase__ : List[Any] = UperNetConvModule(_snake_case ,self.channels ,kernel_size=1 ) lowercase__ : Optional[int] = UperNetConvModule(self.channels ,self.channels ,kernel_size=3 ,padding=1 ) self.lateral_convs.append(_snake_case ) self.fpn_convs.append(_snake_case ) lowercase__ : int = UperNetConvModule( len(self.in_channels ) * self.channels ,self.channels ,kernel_size=3 ,padding=1 ,) def UpperCAmelCase ( self : Dict ) -> Union[str, Any]: """simple docstring""" self.apply(self._init_weights ) def UpperCAmelCase ( self : Union[str, Any] ,_snake_case : Optional[Any] ) -> List[str]: """simple docstring""" if isinstance(_snake_case ,nn.Convad ): module.weight.data.normal_(mean=0.0 ,std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def UpperCAmelCase ( self : Optional[Any] ,_snake_case : Optional[Any] ) -> str: """simple docstring""" lowercase__ : Dict = inputs[-1] lowercase__ : Optional[int] = [x] psp_outs.extend(self.psp_modules(_snake_case ) ) lowercase__ : Optional[Any] = torch.cat(_snake_case ,dim=1 ) lowercase__ : List[str] = self.bottleneck(_snake_case ) return output def UpperCAmelCase ( self : List[str] ,_snake_case : torch.Tensor ) -> torch.Tensor: """simple docstring""" lowercase__ : Tuple = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(_snake_case ) ) # build top-down path lowercase__ : List[Any] = len(_snake_case ) for i in range(used_backbone_levels - 1 ,0 ,-1 ): lowercase__ : Union[str, Any] = laterals[i - 1].shape[2:] lowercase__ : int = laterals[i - 1] + nn.functional.interpolate( laterals[i] ,size=_snake_case ,mode='''bilinear''' ,align_corners=self.align_corners ) # build outputs lowercase__ : List[str] = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 ,0 ,-1 ): lowercase__ : Any = nn.functional.interpolate( fpn_outs[i] ,size=fpn_outs[0].shape[2:] ,mode='''bilinear''' ,align_corners=self.align_corners ) lowercase__ : Any = torch.cat(_snake_case ,dim=1 ) lowercase__ : Any = self.fpn_bottleneck(_snake_case ) lowercase__ : str = self.classifier(_snake_case ) return output class __A ( nn.Module ): '''simple docstring''' def __init__( self : Dict ,_snake_case : List[Any] ,_snake_case : int = 2 ,_snake_case : int = 3 ,_snake_case : Union[int, Tuple[int, int]] = 1 ) -> None: """simple docstring""" super().__init__() lowercase__ : int = config lowercase__ : Dict = config.auxiliary_in_channels lowercase__ : Optional[int] = config.auxiliary_channels lowercase__ : List[Any] = config.auxiliary_num_convs lowercase__ : List[Any] = config.auxiliary_concat_input lowercase__ : str = in_index lowercase__ : Any = (kernel_size // 2) * dilation lowercase__ : Optional[Any] = [] convs.append( UperNetConvModule( self.in_channels ,self.channels ,kernel_size=_snake_case ,padding=_snake_case ,dilation=_snake_case ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels ,self.channels ,kernel_size=_snake_case ,padding=_snake_case ,dilation=_snake_case ) ) if self.num_convs == 0: lowercase__ : List[str] = nn.Identity() else: lowercase__ : Dict = nn.Sequential(*_snake_case ) if self.concat_input: lowercase__ : int = UperNetConvModule( self.in_channels + self.channels ,self.channels ,kernel_size=_snake_case ,padding=kernel_size // 2 ) lowercase__ : List[str] = nn.Convad(self.channels ,config.num_labels ,kernel_size=1 ) def UpperCAmelCase ( self : Optional[Any] ) -> List[str]: """simple docstring""" self.apply(self._init_weights ) def UpperCAmelCase ( self : List[Any] ,_snake_case : List[Any] ) -> Dict: """simple docstring""" if isinstance(_snake_case ,nn.Convad ): module.weight.data.normal_(mean=0.0 ,std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def UpperCAmelCase ( self : List[str] ,_snake_case : torch.Tensor ) -> torch.Tensor: """simple docstring""" lowercase__ : str = encoder_hidden_states[self.in_index] lowercase__ : List[str] = self.convs(_snake_case ) if self.concat_input: lowercase__ : Any = self.conv_cat(torch.cat([hidden_states, output] ,dim=1 ) ) lowercase__ : Dict = self.classifier(_snake_case ) return output class __A ( A_ ): '''simple docstring''' lowerCAmelCase : Any = UperNetConfig lowerCAmelCase : str = "pixel_values" lowerCAmelCase : Dict = True def UpperCAmelCase ( self : int ,_snake_case : str ) -> Optional[int]: """simple docstring""" if isinstance(_snake_case ,_snake_case ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def UpperCAmelCase ( self : List[str] ) -> Dict: """simple docstring""" self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def UpperCAmelCase ( self : int ,_snake_case : str ,_snake_case : str=False ) -> List[str]: """simple docstring""" if isinstance(_snake_case ,_snake_case ): lowercase__ : List[Any] = value lowerCAmelCase_ = R'\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCAmelCase_ = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." ,A_ ,) class __A ( A_ ): '''simple docstring''' def __init__( self : Optional[Any] ,_snake_case : Tuple ) -> int: """simple docstring""" super().__init__(_snake_case ) lowercase__ : int = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) lowercase__ : Any = UperNetHead(_snake_case ,in_channels=self.backbone.channels ) lowercase__ : str = UperNetFCNHead(_snake_case ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('''batch_size, sequence_length''' ) ) @replace_return_docstrings(output_type=_snake_case ,config_class=_CONFIG_FOR_DOC ) def UpperCAmelCase ( self : Dict ,_snake_case : Optional[torch.Tensor] = None ,_snake_case : Optional[bool] = None ,_snake_case : Optional[bool] = None ,_snake_case : Optional[torch.Tensor] = None ,_snake_case : Optional[bool] = None ,) -> Union[tuple, SemanticSegmenterOutput]: """simple docstring""" lowercase__ : int = return_dict if return_dict is not None else self.config.use_return_dict lowercase__ : Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase__ : Any = output_attentions if output_attentions is not None else self.config.output_attentions lowercase__ : Optional[Any] = self.backbone.forward_with_filtered_kwargs( _snake_case ,output_hidden_states=_snake_case ,output_attentions=_snake_case ) lowercase__ : Optional[int] = outputs.feature_maps lowercase__ : Tuple = self.decode_head(_snake_case ) lowercase__ : Optional[int] = nn.functional.interpolate(_snake_case ,size=pixel_values.shape[2:] ,mode='''bilinear''' ,align_corners=_snake_case ) lowercase__ : List[str] = None if self.auxiliary_head is not None: lowercase__ : str = self.auxiliary_head(_snake_case ) lowercase__ : Dict = nn.functional.interpolate( _snake_case ,size=pixel_values.shape[2:] ,mode='''bilinear''' ,align_corners=_snake_case ) lowercase__ : Any = None if labels is not None: if self.config.num_labels == 1: raise ValueError('''The number of labels should be greater than one''' ) else: # compute weighted loss lowercase__ : Union[str, Any] = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) lowercase__ : List[str] = loss_fct(_snake_case ,_snake_case ) lowercase__ : List[str] = loss_fct(_snake_case ,_snake_case ) lowercase__ : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: lowercase__ : Tuple = (logits,) + outputs[1:] else: lowercase__ : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=_snake_case ,logits=_snake_case ,hidden_states=outputs.hidden_states ,attentions=outputs.attentions ,)
302
0
print((lambda quine: quine % quine)("""print((lambda quine: quine %% quine)(%r))"""))
176
from collections.abc import Generator from math import sin def SCREAMING_SNAKE_CASE_ ( __magic_name__ : bytes ) -> bytes: """simple docstring""" if len(__magic_name__ ) != 32: raise ValueError("""Input must be of length 32""" ) UpperCamelCase :int = B"""""" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int ) -> bytes: """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) UpperCamelCase :Any = format(__magic_name__ , """08x""" )[-8:] UpperCamelCase :Union[str, Any] = B"""""" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("""utf-8""" ) return little_endian_hex def SCREAMING_SNAKE_CASE_ ( __magic_name__ : bytes ) -> bytes: """simple docstring""" UpperCamelCase :str = B"""""" for char in message: bit_string += format(__magic_name__ , """08b""" ).encode("""utf-8""" ) UpperCamelCase :Any = format(len(__magic_name__ ) , """064b""" ).encode("""utf-8""" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(__magic_name__ ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def SCREAMING_SNAKE_CASE_ ( __magic_name__ : bytes ) -> Generator[list[int], None, None]: """simple docstring""" if len(__magic_name__ ) % 512 != 0: raise ValueError("""Input must have length that's a multiple of 512""" ) for pos in range(0 , len(__magic_name__ ) , 512 ): UpperCamelCase :Tuple = bit_string[pos : pos + 512] UpperCamelCase :Optional[int] = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int ) -> int: """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) UpperCamelCase :List[str] = format(__magic_name__ , """032b""" ) UpperCamelCase :Any = """""" for c in i_str: new_str += "1" if c == "0" else "0" return int(__magic_name__ , 2 ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int , __magic_name__ : int ) -> int: """simple docstring""" return (a + b) % 2**32 def SCREAMING_SNAKE_CASE_ ( __magic_name__ : int , __magic_name__ : int ) -> int: """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) if shift < 0: raise ValueError("""Shift must be non-negative""" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def SCREAMING_SNAKE_CASE_ ( __magic_name__ : bytes ) -> bytes: """simple docstring""" UpperCamelCase :Tuple = preprocess(__magic_name__ ) UpperCamelCase :List[str] = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states UpperCamelCase :Union[str, Any] = 0X67_45_23_01 UpperCamelCase :Union[str, Any] = 0XEF_CD_AB_89 UpperCamelCase :List[str] = 0X98_BA_DC_FE UpperCamelCase :int = 0X10_32_54_76 UpperCamelCase :int = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(__magic_name__ ): UpperCamelCase :Optional[Any] = aa UpperCamelCase :Any = ba UpperCamelCase :Tuple = ca UpperCamelCase :List[str] = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f UpperCamelCase :int = d ^ (b & (c ^ d)) UpperCamelCase :Optional[int] = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f UpperCamelCase :str = c ^ (d & (b ^ c)) UpperCamelCase :Union[str, Any] = (5 * i + 1) % 16 elif i <= 47: UpperCamelCase :str = b ^ c ^ d UpperCamelCase :Optional[int] = (3 * i + 5) % 16 else: UpperCamelCase :List[str] = c ^ (b | not_aa(__magic_name__ )) UpperCamelCase :int = (7 * i) % 16 UpperCamelCase :Dict = (f + a + added_consts[i] + block_words[g]) % 2**32 UpperCamelCase :Tuple = d UpperCamelCase :str = c UpperCamelCase :Tuple = b UpperCamelCase :Optional[Any] = sum_aa(__magic_name__ , left_rotate_aa(__magic_name__ , shift_amounts[i] ) ) # Add hashed chunk to running total UpperCamelCase :List[str] = sum_aa(__magic_name__ , __magic_name__ ) UpperCamelCase :str = sum_aa(__magic_name__ , __magic_name__ ) UpperCamelCase :int = sum_aa(__magic_name__ , __magic_name__ ) UpperCamelCase :Optional[Any] = sum_aa(__magic_name__ , __magic_name__ ) UpperCamelCase :Optional[Any] = reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) return digest if __name__ == "__main__": import doctest doctest.testmod()
38
0
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCamelCase = """base_with_context""" def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : List[str] = nn.Parameter(torch.FloatTensor(weights['''token_embedder''']['''embedding'''] ) ) A_ : Optional[int] = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=SCREAMING_SNAKE_CASE ) for lyr_num, lyr in enumerate(model.encoders ): A_ : List[Any] = weights[f'''layers_{lyr_num}'''] A_ : Any = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) A_ : List[Any] = ly_weight['''attention'''] A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) A_ : Any = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) A_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) A_ : Any = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) A_ : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) A_ : List[str] = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : Any = nn.Parameter(torch.FloatTensor(weights['''input_proj''']['''kernel'''].T ) ) A_ : Dict = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=SCREAMING_SNAKE_CASE ) for lyr_num, lyr in enumerate(model.encoders ): A_ : List[str] = weights[f'''layers_{lyr_num}'''] A_ : Optional[int] = ly_weight['''attention'''] A_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) A_ : Dict = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) A_ : Any = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) A_ : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) A_ : List[str] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) A_ : List[str] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) A_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : Any = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense0''']['''kernel'''].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense1''']['''kernel'''].T ) ) A_ : List[str] = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=SCREAMING_SNAKE_CASE ) A_ : Dict = nn.Parameter( torch.FloatTensor(weights['''continuous_inputs_projection''']['''kernel'''].T ) ) for lyr_num, lyr in enumerate(model.decoders ): A_ : Any = weights[f'''layers_{lyr_num}'''] A_ : str = nn.Parameter( torch.FloatTensor(ly_weight['''pre_self_attention_layer_norm''']['''scale'''] ) ) A_ : Dict = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_0''']['''DenseGeneral_0''']['''kernel'''].T ) ) A_ : Tuple = ly_weight['''self_attention'''] A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) A_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) A_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) A_ : Optional[Any] = ly_weight['''MultiHeadDotProductAttention_0'''] A_ : int = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) A_ : Dict = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) A_ : Any = nn.Parameter( torch.FloatTensor(ly_weight['''pre_cross_attention_layer_norm''']['''scale'''] ) ) A_ : str = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) A_ : List[str] = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_1''']['''DenseGeneral_0''']['''kernel'''].T ) ) A_ : str = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) A_ : Any = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(weights['''decoder_norm''']['''scale'''] ) ) A_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(weights['''spec_out_dense''']['''kernel'''].T ) ) return model def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): A_ : Tuple = checkpoints.load_tax_checkpoint(args.checkpoint_path ) A_ : Any = jnp.tree_util.tree_map(onp.array , SCREAMING_SNAKE_CASE ) A_ : str = [ '''from __gin__ import dynamic_registration''', '''from music_spectrogram_diffusion.models.diffusion import diffusion_utils''', '''diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0''', '''diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()''', ] A_ : str = os.path.join(args.checkpoint_path , '''..''' , '''config.gin''' ) A_ : Union[str, Any] = inference.parse_training_gin_file(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) A_ : Optional[int] = inference.InferenceModel(args.checkpoint_path , SCREAMING_SNAKE_CASE ) A_ : Tuple = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' , variance_type='''fixed_large''' ) A_ : str = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['''inputs'''] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) A_ : Optional[Any] = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['''targets_context'''] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) A_ : List[Any] = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['''targets_context'''] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) A_ : str = load_notes_encoder(ta_checkpoint['''target''']['''token_encoder'''] , SCREAMING_SNAKE_CASE ) A_ : Optional[int] = load_continuous_encoder(ta_checkpoint['''target''']['''continuous_encoder'''] , SCREAMING_SNAKE_CASE ) A_ : Tuple = load_decoder(ta_checkpoint['''target''']['''decoder'''] , SCREAMING_SNAKE_CASE ) A_ : Any = OnnxRuntimeModel.from_pretrained('''kashif/soundstream_mel_decoder''' ) A_ : Dict = SpectrogramDiffusionPipeline( notes_encoder=SCREAMING_SNAKE_CASE , continuous_encoder=SCREAMING_SNAKE_CASE , decoder=SCREAMING_SNAKE_CASE , scheduler=SCREAMING_SNAKE_CASE , melgan=SCREAMING_SNAKE_CASE , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--output_path""", default=None, type=str, required=True, help="""Path to the converted model.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument( """--checkpoint_path""", default=F'''{MODEL}/checkpoint_500000''', type=str, required=False, help="""Path to the original jax model checkpoint.""", ) UpperCamelCase = parser.parse_args() main(args)
367
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class _lowerCamelCase ( UpperCamelCase ): """simple docstring""" snake_case = ["image_processor"] snake_case = "SamImageProcessor" def __init__( self , _SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' super().__init__(_SCREAMING_SNAKE_CASE ) A_ : Any = self.image_processor A_ : Optional[int] = -10 A_ : List[Any] = self.image_processor.size['''longest_edge'''] def __call__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , )->BatchEncoding: '''simple docstring''' A_ : Union[str, Any] = self.image_processor( _SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) # pop arguments that are not used in the foward but used nevertheless A_ : Tuple = encoding_image_processor['''original_sizes'''] if hasattr(_SCREAMING_SNAKE_CASE , '''numpy''' ): # Checks if Torch or TF tensor A_ : int = original_sizes.numpy() A_ , A_ , A_ : str = self._check_and_preprocess_points( input_points=_SCREAMING_SNAKE_CASE , input_labels=_SCREAMING_SNAKE_CASE , input_boxes=_SCREAMING_SNAKE_CASE , ) A_ : Optional[Any] = self._normalize_and_convert( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , input_points=_SCREAMING_SNAKE_CASE , input_labels=_SCREAMING_SNAKE_CASE , input_boxes=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , ) return encoding_image_processor def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE="pt" , )->Dict: '''simple docstring''' if input_points is not None: if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): A_ : Optional[Any] = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , original_sizes[0] ) for point in input_points ] else: A_ : str = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for point, original_size in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: A_ , A_ : Optional[Any] = self._pad_points_and_labels(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) A_ : List[str] = np.array(_SCREAMING_SNAKE_CASE ) if input_labels is not None: A_ : Dict = np.array(_SCREAMING_SNAKE_CASE ) if input_boxes is not None: if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): A_ : Tuple = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , original_sizes[0] , is_bounding_box=_SCREAMING_SNAKE_CASE ) for box in input_boxes ] else: A_ : List[Any] = [ self._normalize_coordinates(self.target_size , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , is_bounding_box=_SCREAMING_SNAKE_CASE ) for box, original_size in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ] A_ : Union[str, Any] = np.array(_SCREAMING_SNAKE_CASE ) if input_boxes is not None: if return_tensors == "pt": A_ : Dict = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # boxes batch size of 1 by default A_ : Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": A_ : Optional[int] = tf.convert_to_tensor(_SCREAMING_SNAKE_CASE ) # boxes batch size of 1 by default A_ : List[Any] = tf.expand_dims(_SCREAMING_SNAKE_CASE , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'''input_boxes''': input_boxes} ) if input_points is not None: if return_tensors == "pt": A_ : Union[str, Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default A_ : Union[str, Any] = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": A_ : List[str] = tf.convert_to_tensor(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default A_ : Union[str, Any] = tf.expand_dims(_SCREAMING_SNAKE_CASE , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'''input_points''': input_points} ) if input_labels is not None: if return_tensors == "pt": A_ : Optional[Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default A_ : List[Any] = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": A_ : int = tf.convert_to_tensor(_SCREAMING_SNAKE_CASE ) # point batch size of 1 by default A_ : List[Any] = tf.expand_dims(_SCREAMING_SNAKE_CASE , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'''input_labels''': input_labels} ) return encoding_image_processor def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )->Dict: '''simple docstring''' A_ : Optional[Any] = max([point.shape[0] for point in input_points] ) A_ : int = [] for i, point in enumerate(_SCREAMING_SNAKE_CASE ): if point.shape[0] != expected_nb_points: A_ : Optional[int] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) A_ : int = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(_SCREAMING_SNAKE_CASE ) A_ : Optional[int] = processed_input_points return input_points, input_labels def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False )->np.ndarray: '''simple docstring''' A_ , A_ : str = original_size A_ , A_ : Dict = self.image_processor._get_preprocess_shape(_SCREAMING_SNAKE_CASE , longest_edge=_SCREAMING_SNAKE_CASE ) A_ : Optional[int] = deepcopy(_SCREAMING_SNAKE_CASE ).astype(_SCREAMING_SNAKE_CASE ) if is_bounding_box: A_ : Union[str, Any] = coords.reshape(-1 , 2 , 2 ) A_ : Any = coords[..., 0] * (new_w / old_w) A_ : List[str] = coords[..., 1] * (new_h / old_h) if is_bounding_box: A_ : str = coords.reshape(-1 , 4 ) return coords def _snake_case ( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , )->str: '''simple docstring''' if input_points is not None: if hasattr(_SCREAMING_SNAKE_CASE , '''numpy''' ): # Checks for TF or Torch tensor A_ : List[str] = input_points.numpy().tolist() if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not isinstance(input_points[0] , _SCREAMING_SNAKE_CASE ): raise ValueError('''Input points must be a list of list of floating points.''' ) A_ : Optional[Any] = [np.array(_SCREAMING_SNAKE_CASE ) for input_point in input_points] else: A_ : Tuple = None if input_labels is not None: if hasattr(_SCREAMING_SNAKE_CASE , '''numpy''' ): A_ : Dict = input_labels.numpy().tolist() if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not isinstance(input_labels[0] , _SCREAMING_SNAKE_CASE ): raise ValueError('''Input labels must be a list of list integers.''' ) A_ : Union[str, Any] = [np.array(_SCREAMING_SNAKE_CASE ) for label in input_labels] else: A_ : str = None if input_boxes is not None: if hasattr(_SCREAMING_SNAKE_CASE , '''numpy''' ): A_ : str = input_boxes.numpy().tolist() if ( not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not isinstance(input_boxes[0] , _SCREAMING_SNAKE_CASE ) or not isinstance(input_boxes[0][0] , _SCREAMING_SNAKE_CASE ) ): raise ValueError('''Input boxes must be a list of list of list of floating points.''' ) A_ : Tuple = [np.array(_SCREAMING_SNAKE_CASE ).astype(np.floataa ) for box in input_boxes] else: A_ : Dict = None return input_points, input_labels, input_boxes @property def _snake_case ( self )->List[str]: '''simple docstring''' A_ : Optional[Any] = self.image_processor.model_input_names return list(dict.fromkeys(_SCREAMING_SNAKE_CASE ) ) def _snake_case ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' return self.image_processor.post_process_masks(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )
65
0
'''simple docstring''' import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml a_ : List[str] = NewType("DataClass", Any) a_ : List[str] = NewType("DataClassType", Any) def _A (lowerCAmelCase__ :List[str] ) -> List[str]: '''simple docstring''' if isinstance(snake_case_ , snake_case_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f'Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).' ) def _A (lowerCAmelCase__ :list ) -> Callable[[str], Any]: '''simple docstring''' _a = {str(snake_case_ ): choice for choice in choices} return lambda lowerCAmelCase__ : str_to_choice.get(snake_case_ , snake_case_ ) def _A (*, lowerCAmelCase__ :Union[str, List[str]] = None , lowerCAmelCase__ :str = None , lowerCAmelCase__ :Any = dataclasses.MISSING , lowerCAmelCase__ :Callable[[], Any] = dataclasses.MISSING , lowerCAmelCase__ :dict = None , **lowerCAmelCase__ :str , ) -> dataclasses.Field: '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls _a = {} if aliases is not None: _a = aliases if help is not None: _a = help return dataclasses.field(metadata=snake_case_ , default=snake_case_ , default_factory=snake_case_ , **snake_case_ ) class a ( SCREAMING_SNAKE_CASE__ ): _lowerCAmelCase = 4_2 def __init__( self , __magic_name__ , **__magic_name__ ) -> List[Any]: if "formatter_class" not in kwargs: _a = ArgumentDefaultsHelpFormatter super().__init__(**__lowerCamelCase ) if dataclasses.is_dataclass(__lowerCamelCase ): _a = [dataclass_types] _a = list(__lowerCamelCase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__lowerCamelCase ) @staticmethod def __UpperCAmelCase ( __magic_name__ , __magic_name__ ) -> Optional[Any]: _a = f'--{field.name}' _a = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __lowerCamelCase ): raise RuntimeError( 'Unresolved type detected, which should have been done with the help of ' '`typing.get_type_hints` method by default' ) _a = kwargs.pop('aliases' , [] ) if isinstance(__lowerCamelCase , __lowerCamelCase ): _a = [aliases] _a = getattr(field.type , '__origin__' , field.type ) if origin_type is Union or (hasattr(__lowerCamelCase , 'UnionType' ) and isinstance(__lowerCamelCase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__lowerCamelCase ) not in field.type.__args__ ): raise ValueError( 'Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because' ' the argument parser only supports one type per argument.' f' Problem encountered in field \'{field.name}\'.' ) if type(__lowerCamelCase ) not in field.type.__args__: # filter `str` in Union _a = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] _a = getattr(field.type , '__origin__' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) _a = ( field.type.__args__[0] if isinstance(__lowerCamelCase , field.type.__args__[1] ) else field.type.__args__[1] ) _a = getattr(field.type , '__origin__' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) _a = {} if origin_type is Literal or (isinstance(field.type , __lowerCamelCase ) and issubclass(field.type , __lowerCamelCase )): if origin_type is Literal: _a = field.type.__args__ else: _a = [x.value for x in field.type] _a = make_choice_type_function(kwargs['choices'] ) if field.default is not dataclasses.MISSING: _a = field.default else: _a = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument _a = copy(__lowerCamelCase ) # Hack because type=bool in argparse does not behave as we want. _a = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. _a = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way _a = default # This tells argparse we accept 0 or 1 value after --field_name _a = '''?''' # This is the value that will get picked if we do --field_name (without value) _a = True elif isclass(__lowerCamelCase ) and issubclass(__lowerCamelCase , __lowerCamelCase ): _a = field.type.__args__[0] _a = '''+''' if field.default_factory is not dataclasses.MISSING: _a = field.default_factory() elif field.default is dataclasses.MISSING: _a = True else: _a = field.type if field.default is not dataclasses.MISSING: _a = field.default elif field.default_factory is not dataclasses.MISSING: _a = field.default_factory() else: _a = True parser.add_argument(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): _a = False parser.add_argument(f'--no_{field.name}' , action='store_false' , dest=field.name , **__lowerCamelCase ) def __UpperCAmelCase ( self , __magic_name__ ) -> List[str]: if hasattr(__lowerCamelCase , '_argument_group_name' ): _a = self.add_argument_group(dtype._argument_group_name ) else: _a = self try: _a = get_type_hints(__lowerCamelCase ) except NameError: raise RuntimeError( f'Type resolution failed for {dtype}. Try declaring the class in global scope or ' 'removing line of `from __future__ import annotations` which opts in Postponed ' 'Evaluation of Annotations (PEP 563)' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(__lowerCamelCase ): _a = '''.'''.join(map(__lowerCamelCase , sys.version_info[:3] ) ) raise RuntimeError( f'Type resolution failed for {dtype} on Python {python_version}. Try removing ' 'line of `from __future__ import annotations` which opts in union types as ' '`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ' 'support Python versions that lower than 3.10, you need to use ' '`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ' '`X | None`.' ) from ex raise for field in dataclasses.fields(__lowerCamelCase ): if not field.init: continue _a = type_hints[field.name] self._parse_dataclass_field(__lowerCamelCase , __lowerCamelCase ) def __UpperCAmelCase ( self , __magic_name__=None , __magic_name__=False , __magic_name__=True , __magic_name__=None , __magic_name__=None , ) -> Optional[Any]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): _a = [] if args_filename: args_files.append(Path(__lowerCamelCase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('.args' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values _a = ArgumentParser() args_file_parser.add_argument(__lowerCamelCase , type=__lowerCamelCase , action='append' ) # Use only remaining args for further parsing (remove the args_file_flag) _a = args_file_parser.parse_known_args(args=__lowerCamelCase ) _a = vars(__lowerCamelCase ).get(args_file_flag.lstrip('-' ) , __lowerCamelCase ) if cmd_args_file_paths: args_files.extend([Path(__lowerCamelCase ) for p in cmd_args_file_paths] ) _a = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last _a = file_args + args if args is not None else file_args + sys.argv[1:] _a = self.parse_known_args(args=__lowerCamelCase ) _a = [] for dtype in self.dataclass_types: _a = {f.name for f in dataclasses.fields(__lowerCamelCase ) if f.init} _a = {k: v for k, v in vars(__lowerCamelCase ).items() if k in keys} for k in keys: delattr(__lowerCamelCase , __lowerCamelCase ) _a = dtype(**__lowerCamelCase ) outputs.append(__lowerCamelCase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__lowerCamelCase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'Some specified arguments are not used by the HfArgumentParser: {remaining_args}' ) return (*outputs,) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = False ) -> Union[str, Any]: _a = set(args.keys() ) _a = [] for dtype in self.dataclass_types: _a = {f.name for f in dataclasses.fields(__lowerCamelCase ) if f.init} _a = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) _a = dtype(**__lowerCamelCase ) outputs.append(__lowerCamelCase ) if not allow_extra_keys and unused_keys: raise ValueError(f'Some keys are not used by the HfArgumentParser: {sorted(__lowerCamelCase )}' ) return tuple(__lowerCamelCase ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = False ) -> str: with open(Path(__lowerCamelCase ) , encoding='utf-8' ) as open_json_file: _a = json.loads(open_json_file.read() ) _a = self.parse_dict(__lowerCamelCase , allow_extra_keys=__lowerCamelCase ) return tuple(__lowerCamelCase ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = False ) -> Any: _a = self.parse_dict(yaml.safe_load(Path(__lowerCamelCase ).read_text() ) , allow_extra_keys=__lowerCamelCase ) return tuple(__lowerCamelCase )
168
"""simple docstring""" import logging from transformers.configuration_utils import PretrainedConfig a_ = logging.getLogger(__name__) class __snake_case ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = """masked_bert""" def __init__( self , __lowerCamelCase=3_0522 , __lowerCamelCase=768 , __lowerCamelCase=12 , __lowerCamelCase=12 , __lowerCamelCase=3072 , __lowerCamelCase="gelu" , __lowerCamelCase=0.1 , __lowerCamelCase=0.1 , __lowerCamelCase=512 , __lowerCamelCase=2 , __lowerCamelCase=0.0_2 , __lowerCamelCase=1e-1_2 , __lowerCamelCase=0 , __lowerCamelCase="topK" , __lowerCamelCase="constant" , __lowerCamelCase=0.0 , **__lowerCamelCase , ): '''simple docstring''' super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) __A : Dict = vocab_size __A : Union[str, Any] = hidden_size __A : Tuple = num_hidden_layers __A : Tuple = num_attention_heads __A : Optional[Any] = hidden_act __A : List[str] = intermediate_size __A : Any = hidden_dropout_prob __A : Optional[Any] = attention_probs_dropout_prob __A : Any = max_position_embeddings __A : str = type_vocab_size __A : List[Any] = initializer_range __A : str = layer_norm_eps __A : Optional[int] = pruning_method __A : str = mask_init __A : Any = mask_scale
179
0
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCAmelCase__ = logging.get_logger(__name__) class snake_case__(enum.Enum ): """simple docstring""" lowercase_ = 0 lowercase_ = 1 @add_end_docstrings(__UpperCamelCase ) class snake_case__(__UpperCamelCase ): """simple docstring""" lowercase_ = "generated" def __init__( self : Optional[int] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Dict ): super().__init__(*_lowerCAmelCase , **_lowerCAmelCase ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def snake_case ( self : int , SCREAMING_SNAKE_CASE : Any=None , SCREAMING_SNAKE_CASE : Any=None , SCREAMING_SNAKE_CASE : List[Any]=None , SCREAMING_SNAKE_CASE : Optional[Any]=None , SCREAMING_SNAKE_CASE : Optional[int]=None , SCREAMING_SNAKE_CASE : Tuple=None , **SCREAMING_SNAKE_CASE : Dict , ): lowercase__ : Tuple = {} if truncation is not None: lowercase__ : List[Any] = truncation lowercase__ : Any = generate_kwargs lowercase__ : Union[str, Any] = {} if return_tensors is not None and return_type is None: lowercase__ : List[str] = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: lowercase__ : int = return_type if clean_up_tokenization_spaces is not None: lowercase__ : List[Any] = clean_up_tokenization_spaces if stop_sequence is not None: lowercase__ : Optional[Any] = self.tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) if len(_lowerCAmelCase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) lowercase__ : Tuple = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def snake_case ( self : Union[str, Any] , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): return True def snake_case ( self : List[str] , *SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Optional[int] ): lowercase__ : Any = self.model.config.prefix if self.model.config.prefix is not None else """""" if isinstance(args[0] , _lowerCAmelCase ): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input" ) lowercase__ : Optional[int] = ([prefix + arg for arg in args[0]],) lowercase__ : List[Any] = True elif isinstance(args[0] , _lowerCAmelCase ): lowercase__ : Tuple = (prefix + args[0],) lowercase__ : Tuple = False else: raise ValueError( f""" `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`""" ) lowercase__ : Dict = self.tokenizer(*_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Optional[int] , **SCREAMING_SNAKE_CASE : str ): lowercase__ : Optional[Any] = super().__call__(*_lowerCAmelCase , **_lowerCAmelCase ) if ( isinstance(args[0] , _lowerCAmelCase ) and all(isinstance(_lowerCAmelCase , _lowerCAmelCase ) for el in args[0] ) and all(len(_lowerCAmelCase ) == 1 for res in result ) ): return [res[0] for res in result] return result def snake_case ( self : str , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Any=TruncationStrategy.DO_NOT_TRUNCATE , **SCREAMING_SNAKE_CASE : Dict ): lowercase__ : str = self._parse_and_tokenize(_lowerCAmelCase , truncation=_lowerCAmelCase , **_lowerCAmelCase ) return inputs def snake_case ( self : Union[str, Any] , SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): if self.framework == "pt": lowercase__ : Dict = model_inputs["""input_ids"""].shape elif self.framework == "tf": lowercase__ : Union[str, Any] = tf.shape(model_inputs["input_ids"] ).numpy() lowercase__ : Optional[Any] = generate_kwargs.get("min_length" , self.model.config.min_length ) lowercase__ : Optional[Any] = generate_kwargs.get("max_length" , self.model.config.max_length ) self.check_inputs(_lowerCAmelCase , generate_kwargs["min_length"] , generate_kwargs["max_length"] ) lowercase__ : Tuple = self.model.generate(**_lowerCAmelCase , **_lowerCAmelCase ) lowercase__ : Dict = output_ids.shape[0] if self.framework == "pt": lowercase__ : Union[str, Any] = output_ids.reshape(_lowerCAmelCase , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": lowercase__ : Any = tf.reshape(_lowerCAmelCase , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def snake_case ( self : Optional[Any] , SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : str=ReturnType.TEXT , SCREAMING_SNAKE_CASE : Dict=False ): lowercase__ : Union[str, Any] = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: lowercase__ : Any = {f"""{self.return_name}_token_ids""": output_ids} elif return_type == ReturnType.TEXT: lowercase__ : List[str] = { f"""{self.return_name}_text""": self.tokenizer.decode( _lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase , ) } records.append(_lowerCAmelCase ) return records @add_end_docstrings(__UpperCamelCase ) class snake_case__(__UpperCamelCase ): """simple docstring""" lowercase_ = "summary" def __call__( self : Optional[int] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Dict ): return super().__call__(*_lowerCAmelCase , **_lowerCAmelCase ) def snake_case ( self : Tuple , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): if max_length < min_length: logger.warning(f"""Your min_length={min_length} must be inferior than your max_length={max_length}.""" ) if input_length < max_length: logger.warning( f"""Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is """ "a summarization task, where outputs shorter than the input are typically wanted, you might " f"""consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})""" ) @add_end_docstrings(__UpperCamelCase ) class snake_case__(__UpperCamelCase ): """simple docstring""" lowercase_ = "translation" def snake_case ( self : List[str] , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): if input_length > 0.9 * max_length: logger.warning( f"""Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider """ "increasing your max_length manually, e.g. translator('...', max_length=400)" ) return True def snake_case ( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Union[str, Any]=TruncationStrategy.DO_NOT_TRUNCATE , SCREAMING_SNAKE_CASE : List[str]=None , SCREAMING_SNAKE_CASE : List[Any]=None ): if getattr(self.tokenizer , "_build_translation_inputs" , _lowerCAmelCase ): return self.tokenizer._build_translation_inputs( *_lowerCAmelCase , return_tensors=self.framework , truncation=_lowerCAmelCase , src_lang=_lowerCAmelCase , tgt_lang=_lowerCAmelCase ) else: return super()._parse_and_tokenize(*_lowerCAmelCase , truncation=_lowerCAmelCase ) def snake_case ( self : Dict , SCREAMING_SNAKE_CASE : Any=None , SCREAMING_SNAKE_CASE : Optional[Any]=None , **SCREAMING_SNAKE_CASE : Union[str, Any] ): lowercase__ : Optional[int] = super()._sanitize_parameters(**_lowerCAmelCase ) if src_lang is not None: lowercase__ : List[str] = src_lang if tgt_lang is not None: lowercase__ : str = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. lowercase__ : Optional[Any] = kwargs.get("task" , self.task ) lowercase__ : List[Any] = task.split("_" ) if task and len(_lowerCAmelCase ) == 4: # translation, XX, to YY lowercase__ : str = items[1] lowercase__ : str = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : Dict , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Dict ): return super().__call__(*_lowerCAmelCase , **_lowerCAmelCase )
370
import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__=1 ): """simple docstring""" if n_shave_prefix_segments >= 0: return ".".join(path.split("." )[n_shave_prefix_segments:] ) else: return ".".join(path.split("." )[:n_shave_prefix_segments] ) def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" lowercase__ : List[str] = [] for old_item in old_list: lowercase__ : Optional[Any] = old_item.replace("in_layers.0" , "norm1" ) lowercase__ : Union[str, Any] = new_item.replace("in_layers.2" , "conv1" ) lowercase__ : Optional[Any] = new_item.replace("out_layers.0" , "norm2" ) lowercase__ : Union[str, Any] = new_item.replace("out_layers.3" , "conv2" ) lowercase__ : Dict = new_item.replace("emb_layers.1" , "time_emb_proj" ) lowercase__ : int = new_item.replace("skip_connection" , "conv_shortcut" ) lowercase__ : Tuple = shave_segments(lowerCamelCase__ , n_shave_prefix_segments=lowerCamelCase__ ) mapping.append({"old": old_item, "new": new_item} ) return mapping def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" lowercase__ : str = [] for old_item in old_list: lowercase__ : Optional[int] = old_item lowercase__ : Dict = new_item.replace("norm.weight" , "group_norm.weight" ) lowercase__ : Optional[int] = new_item.replace("norm.bias" , "group_norm.bias" ) lowercase__ : Tuple = new_item.replace("proj_out.weight" , "proj_attn.weight" ) lowercase__ : List[Any] = new_item.replace("proj_out.bias" , "proj_attn.bias" ) lowercase__ : Optional[Any] = shave_segments(lowerCamelCase__ , n_shave_prefix_segments=lowerCamelCase__ ) mapping.append({"old": old_item, "new": new_item} ) return mapping def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None ): """simple docstring""" assert isinstance(lowerCamelCase__ , lowerCamelCase__ ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): lowercase__ : List[str] = old_checkpoint[path] lowercase__ : str = old_tensor.shape[0] // 3 lowercase__ : List[str] = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) lowercase__ : Union[str, Any] = old_tensor.shape[0] // config["num_head_channels"] // 3 lowercase__ : Union[str, Any] = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) lowercase__ , lowercase__ , lowercase__ : Optional[Any] = old_tensor.split(channels // num_heads , dim=1 ) lowercase__ : Dict = query.reshape(lowerCamelCase__ ) lowercase__ : Dict = key.reshape(lowerCamelCase__ ) lowercase__ : int = value.reshape(lowerCamelCase__ ) for path in paths: lowercase__ : Union[str, Any] = path["new"] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here lowercase__ : List[Any] = new_path.replace("middle_block.0" , "mid_block.resnets.0" ) lowercase__ : Optional[Any] = new_path.replace("middle_block.1" , "mid_block.attentions.0" ) lowercase__ : List[str] = new_path.replace("middle_block.2" , "mid_block.resnets.1" ) if additional_replacements is not None: for replacement in additional_replacements: lowercase__ : Tuple = new_path.replace(replacement["old"] , replacement["new"] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: lowercase__ : List[Any] = old_checkpoint[path["old"]][:, :, 0] else: lowercase__ : List[Any] = old_checkpoint[path["old"]] def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" lowercase__ : Union[str, Any] = {} lowercase__ : Optional[Any] = checkpoint["time_embed.0.weight"] lowercase__ : Tuple = checkpoint["time_embed.0.bias"] lowercase__ : Dict = checkpoint["time_embed.2.weight"] lowercase__ : Optional[Any] = checkpoint["time_embed.2.bias"] lowercase__ : Optional[int] = checkpoint["input_blocks.0.0.weight"] lowercase__ : List[Any] = checkpoint["input_blocks.0.0.bias"] lowercase__ : Tuple = checkpoint["out.0.weight"] lowercase__ : List[Any] = checkpoint["out.0.bias"] lowercase__ : Tuple = checkpoint["out.2.weight"] lowercase__ : Optional[Any] = checkpoint["out.2.bias"] # Retrieves the keys for the input blocks only lowercase__ : Dict = len({".".join(layer.split("." )[:2] ) for layer in checkpoint if "input_blocks" in layer} ) lowercase__ : str = { layer_id: [key for key in checkpoint if F"""input_blocks.{layer_id}""" in key] for layer_id in range(lowerCamelCase__ ) } # Retrieves the keys for the middle blocks only lowercase__ : Tuple = len({".".join(layer.split("." )[:2] ) for layer in checkpoint if "middle_block" in layer} ) lowercase__ : Union[str, Any] = { layer_id: [key for key in checkpoint if F"""middle_block.{layer_id}""" in key] for layer_id in range(lowerCamelCase__ ) } # Retrieves the keys for the output blocks only lowercase__ : Tuple = len({".".join(layer.split("." )[:2] ) for layer in checkpoint if "output_blocks" in layer} ) lowercase__ : Tuple = { layer_id: [key for key in checkpoint if F"""output_blocks.{layer_id}""" in key] for layer_id in range(lowerCamelCase__ ) } for i in range(1 , lowerCamelCase__ ): lowercase__ : Tuple = (i - 1) // (config["num_res_blocks"] + 1) lowercase__ : Optional[int] = (i - 1) % (config["num_res_blocks"] + 1) lowercase__ : List[Any] = [key for key in input_blocks[i] if F"""input_blocks.{i}.0""" in key] lowercase__ : Dict = [key for key in input_blocks[i] if F"""input_blocks.{i}.1""" in key] if F"""input_blocks.{i}.0.op.weight""" in checkpoint: lowercase__ : int = checkpoint[ F"""input_blocks.{i}.0.op.weight""" ] lowercase__ : List[str] = checkpoint[ F"""input_blocks.{i}.0.op.bias""" ] continue lowercase__ : Union[str, Any] = renew_resnet_paths(lowerCamelCase__ ) lowercase__ : Optional[int] = {"old": F"""input_blocks.{i}.0""", "new": F"""down_blocks.{block_id}.resnets.{layer_in_block_id}"""} lowercase__ : Optional[int] = {"old": "resnets.2.op", "new": "downsamplers.0.op"} assign_to_checkpoint( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , additional_replacements=[meta_path, resnet_op] , config=lowerCamelCase__ ) if len(lowerCamelCase__ ): lowercase__ : Tuple = renew_attention_paths(lowerCamelCase__ ) lowercase__ : str = { "old": F"""input_blocks.{i}.1""", "new": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}""", } lowercase__ : List[str] = { F"""input_blocks.{i}.1.qkv.bias""": { "key": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", "query": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", "value": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""input_blocks.{i}.1.qkv.weight""": { "key": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", "query": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", "value": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , additional_replacements=[meta_path] , attention_paths_to_split=lowerCamelCase__ , config=lowerCamelCase__ , ) lowercase__ : int = middle_blocks[0] lowercase__ : Dict = middle_blocks[1] lowercase__ : Dict = middle_blocks[2] lowercase__ : Any = renew_resnet_paths(lowerCamelCase__ ) assign_to_checkpoint(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , config=lowerCamelCase__ ) lowercase__ : List[Any] = renew_resnet_paths(lowerCamelCase__ ) assign_to_checkpoint(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , config=lowerCamelCase__ ) lowercase__ : Optional[int] = renew_attention_paths(lowerCamelCase__ ) lowercase__ : Optional[int] = { "middle_block.1.qkv.bias": { "key": "mid_block.attentions.0.key.bias", "query": "mid_block.attentions.0.query.bias", "value": "mid_block.attentions.0.value.bias", }, "middle_block.1.qkv.weight": { "key": "mid_block.attentions.0.key.weight", "query": "mid_block.attentions.0.query.weight", "value": "mid_block.attentions.0.value.weight", }, } assign_to_checkpoint( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , attention_paths_to_split=lowerCamelCase__ , config=lowerCamelCase__ ) for i in range(lowerCamelCase__ ): lowercase__ : List[Any] = i // (config["num_res_blocks"] + 1) lowercase__ : Optional[int] = i % (config["num_res_blocks"] + 1) lowercase__ : List[Any] = [shave_segments(lowerCamelCase__ , 2 ) for name in output_blocks[i]] lowercase__ : Optional[Any] = {} for layer in output_block_layers: lowercase__ , lowercase__ : str = layer.split("." )[0], shave_segments(lowerCamelCase__ , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(lowerCamelCase__ ) else: lowercase__ : Tuple = [layer_name] if len(lowerCamelCase__ ) > 1: lowercase__ : Dict = [key for key in output_blocks[i] if F"""output_blocks.{i}.0""" in key] lowercase__ : Dict = [key for key in output_blocks[i] if F"""output_blocks.{i}.1""" in key] lowercase__ : Optional[Any] = renew_resnet_paths(lowerCamelCase__ ) lowercase__ : Optional[Any] = renew_resnet_paths(lowerCamelCase__ ) lowercase__ : Tuple = {"old": F"""output_blocks.{i}.0""", "new": F"""up_blocks.{block_id}.resnets.{layer_in_block_id}"""} assign_to_checkpoint(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , additional_replacements=[meta_path] , config=lowerCamelCase__ ) if ["conv.weight", "conv.bias"] in output_block_list.values(): lowercase__ : List[str] = list(output_block_list.values() ).index(["conv.weight", "conv.bias"] ) lowercase__ : Tuple = checkpoint[ F"""output_blocks.{i}.{index}.conv.weight""" ] lowercase__ : Optional[Any] = checkpoint[ F"""output_blocks.{i}.{index}.conv.bias""" ] # Clear attentions as they have been attributed above. if len(lowerCamelCase__ ) == 2: lowercase__ : int = [] if len(lowerCamelCase__ ): lowercase__ : Tuple = renew_attention_paths(lowerCamelCase__ ) lowercase__ : str = { "old": F"""output_blocks.{i}.1""", "new": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}""", } lowercase__ : Union[str, Any] = { F"""output_blocks.{i}.1.qkv.bias""": { "key": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", "query": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", "value": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""output_blocks.{i}.1.qkv.weight""": { "key": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", "query": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", "value": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any("qkv" in key for key in attentions ) else None , config=lowerCamelCase__ , ) else: lowercase__ : int = renew_resnet_paths(lowerCamelCase__ , n_shave_prefix_segments=1 ) for path in resnet_0_paths: lowercase__ : List[Any] = ".".join(["output_blocks", str(lowerCamelCase__ ), path["old"]] ) lowercase__ : Any = ".".join(["up_blocks", str(lowerCamelCase__ ), "resnets", str(lowerCamelCase__ ), path["new"]] ) lowercase__ : List[Any] = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = torch.load(args.checkpoint_path) with open(args.config_file) as f: lowerCAmelCase__ = json.loads(f.read()) lowerCAmelCase__ = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] lowerCAmelCase__ = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: lowerCAmelCase__ = DDPMScheduler.from_config('''/'''.join(args.checkpoint_path.split('''/''')[:-1])) lowerCAmelCase__ = VQModel.from_pretrained('''/'''.join(args.checkpoint_path.split('''/''')[:-1])) lowerCAmelCase__ = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
121
0
"""simple docstring""" import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :int = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) __SCREAMING_SNAKE_CASE :Dict = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = -1 __SCREAMING_SNAKE_CASE :Dict = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = model.generate(SCREAMING_SNAKE_CASE__ ,max_new_tokens=10 ,do_sample=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: __SCREAMING_SNAKE_CASE :str = TextStreamer(SCREAMING_SNAKE_CASE__ ) model.generate(SCREAMING_SNAKE_CASE__ ,max_new_tokens=10 ,do_sample=SCREAMING_SNAKE_CASE__ ,streamer=SCREAMING_SNAKE_CASE__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __SCREAMING_SNAKE_CASE :Optional[Any] = cs.out[:-1] self.assertEqual(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) __SCREAMING_SNAKE_CASE :str = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[Any] = -1 __SCREAMING_SNAKE_CASE :Dict = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = model.generate(SCREAMING_SNAKE_CASE__ ,max_new_tokens=10 ,do_sample=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = tokenizer.decode(greedy_ids[0] ) __SCREAMING_SNAKE_CASE :Optional[int] = TextIteratorStreamer(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = {'''input_ids''': input_ids, '''max_new_tokens''': 10, '''do_sample''': False, '''streamer''': streamer} __SCREAMING_SNAKE_CASE :int = Thread(target=model.generate ,kwargs=SCREAMING_SNAKE_CASE__ ) thread.start() __SCREAMING_SNAKE_CASE :str = '''''' for new_text in streamer: streamer_text += new_text self.assertEqual(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :str = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) __SCREAMING_SNAKE_CASE :List[str] = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = -1 __SCREAMING_SNAKE_CASE :Dict = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = model.generate(SCREAMING_SNAKE_CASE__ ,max_new_tokens=10 ,do_sample=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[str] = greedy_ids[:, input_ids.shape[1] :] __SCREAMING_SNAKE_CASE :List[Any] = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: __SCREAMING_SNAKE_CASE :Dict = TextStreamer(SCREAMING_SNAKE_CASE__ ,skip_prompt=SCREAMING_SNAKE_CASE__ ) model.generate(SCREAMING_SNAKE_CASE__ ,max_new_tokens=10 ,do_sample=SCREAMING_SNAKE_CASE__ ,streamer=SCREAMING_SNAKE_CASE__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __SCREAMING_SNAKE_CASE :str = cs.out[:-1] self.assertEqual(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = AutoTokenizer.from_pretrained('''distilgpt2''' ) __SCREAMING_SNAKE_CASE :List[str] = AutoModelForCausalLM.from_pretrained('''distilgpt2''' ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = -1 __SCREAMING_SNAKE_CASE :Any = torch.ones((1, 5) ,device=SCREAMING_SNAKE_CASE__ ).long() * model.config.bos_token_id with CaptureStdout() as cs: __SCREAMING_SNAKE_CASE :List[Any] = TextStreamer(SCREAMING_SNAKE_CASE__ ,skip_special_tokens=SCREAMING_SNAKE_CASE__ ) model.generate(SCREAMING_SNAKE_CASE__ ,max_new_tokens=1 ,do_sample=SCREAMING_SNAKE_CASE__ ,streamer=SCREAMING_SNAKE_CASE__ ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token __SCREAMING_SNAKE_CASE :Tuple = cs.out[:-1] # Remove the final "\n" __SCREAMING_SNAKE_CASE :Optional[int] = tokenizer(SCREAMING_SNAKE_CASE__ ,return_tensors='''pt''' ) self.assertEqual(streamer_text_tokenized.input_ids.shape ,(1, 1) ) def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) __SCREAMING_SNAKE_CASE :List[Any] = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = -1 __SCREAMING_SNAKE_CASE :Optional[Any] = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = TextIteratorStreamer(SCREAMING_SNAKE_CASE__ ,timeout=0.0_0_1 ) __SCREAMING_SNAKE_CASE :int = {'''input_ids''': input_ids, '''max_new_tokens''': 10, '''do_sample''': False, '''streamer''': streamer} __SCREAMING_SNAKE_CASE :List[Any] = Thread(target=model.generate ,kwargs=SCREAMING_SNAKE_CASE__ ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(SCREAMING_SNAKE_CASE__ ): __SCREAMING_SNAKE_CASE :Tuple = '''''' for new_text in streamer: streamer_text += new_text
191
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "facebook/xmod-base": "https://huggingface.co/facebook/xmod-base/resolve/main/config.json", "facebook/xmod-large-prenorm": "https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json", "facebook/xmod-base-13-125k": "https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json", "facebook/xmod-base-30-125k": "https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json", "facebook/xmod-base-30-195k": "https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json", "facebook/xmod-base-60-125k": "https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json", "facebook/xmod-base-60-265k": "https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json", "facebook/xmod-base-75-125k": "https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json", "facebook/xmod-base-75-269k": "https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json", } class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : Any = '''xmod''' def __init__( self ,SCREAMING_SNAKE_CASE__=3_05_22 ,SCREAMING_SNAKE_CASE__=7_68 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=30_72 ,SCREAMING_SNAKE_CASE__="gelu" ,SCREAMING_SNAKE_CASE__=0.1 ,SCREAMING_SNAKE_CASE__=0.1 ,SCREAMING_SNAKE_CASE__=5_12 ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=1E-12 ,SCREAMING_SNAKE_CASE__=1 ,SCREAMING_SNAKE_CASE__=0 ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__="absolute" ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=None ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=True ,SCREAMING_SNAKE_CASE__=("en_XX",) ,SCREAMING_SNAKE_CASE__=None ,**SCREAMING_SNAKE_CASE__ ,) -> Optional[Any]: """simple docstring""" super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ ,bos_token_id=SCREAMING_SNAKE_CASE__ ,eos_token_id=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = vocab_size __SCREAMING_SNAKE_CASE :List[Any] = hidden_size __SCREAMING_SNAKE_CASE :List[str] = num_hidden_layers __SCREAMING_SNAKE_CASE :List[str] = num_attention_heads __SCREAMING_SNAKE_CASE :Optional[int] = hidden_act __SCREAMING_SNAKE_CASE :Tuple = intermediate_size __SCREAMING_SNAKE_CASE :Dict = hidden_dropout_prob __SCREAMING_SNAKE_CASE :str = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :Optional[Any] = max_position_embeddings __SCREAMING_SNAKE_CASE :Optional[Any] = type_vocab_size __SCREAMING_SNAKE_CASE :str = initializer_range __SCREAMING_SNAKE_CASE :List[Any] = layer_norm_eps __SCREAMING_SNAKE_CASE :Optional[Any] = position_embedding_type __SCREAMING_SNAKE_CASE :Any = use_cache __SCREAMING_SNAKE_CASE :List[str] = classifier_dropout __SCREAMING_SNAKE_CASE :Any = pre_norm __SCREAMING_SNAKE_CASE :Dict = adapter_reduction_factor __SCREAMING_SNAKE_CASE :Dict = adapter_layer_norm __SCREAMING_SNAKE_CASE :Dict = adapter_reuse_layer_norm __SCREAMING_SNAKE_CASE :Tuple = ln_before_adapter __SCREAMING_SNAKE_CASE :Any = list(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = default_language class _SCREAMING_SNAKE_CASE( A ): @property def _UpperCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": __SCREAMING_SNAKE_CASE :Tuple = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __SCREAMING_SNAKE_CASE :Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
191
1
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument('''--user''', type=str, default='''ubuntu''') parser.add_argument('''--host''', type=str, default='''localhost''') parser.add_argument('''--key_path''', type=str, default=None) parser.add_argument('''--instance''', type=str, default='''V100:1''') parser.add_argument('''--provider''', type=str, default='''cheapest''') parser.add_argument('''--use_spot''', type=bool, default=False) parser.add_argument('''--example''', type=str, default='''pytorch/text-generation/run_generation.py''') lowerCamelCase__,lowerCamelCase__ = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('''Cannot specify both BYO and on-demand cluster args''') lowerCamelCase__ = rh.cluster( name='''rh-cluster''', ips=[args.host], ssh_creds={'''ssh_user''': args.user, '''ssh_private_key''': args.key_path} ) else: lowerCamelCase__ = rh.cluster( name='''rh-cluster''', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) lowerCamelCase__ = args.example.rsplit('''/''', 1)[0] # Set up remote environment cluster.install_packages(['''pip:./''']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F"pip install -r transformers/examples/{example_dir}/requirements.txt"]) cluster.run(['''pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117''']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F"python transformers/examples/{args.example} {' '.join(shlex.quote(arg) for arg in unknown)}"]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
63
import os def lowerCAmelCase__ ( ) ->Any: '''simple docstring''' with open(os.path.dirname(a__ ) + "/grid.txt" ) as f: _UpperCamelCase = [] # noqa: E741 for _ in range(20 ): l.append([int(a__ ) for x in f.readline().split()] ) _UpperCamelCase = 0 # right for i in range(20 ): for j in range(17 ): _UpperCamelCase = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: _UpperCamelCase = temp # down for i in range(17 ): for j in range(20 ): _UpperCamelCase = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: _UpperCamelCase = temp # diagonal 1 for i in range(17 ): for j in range(17 ): _UpperCamelCase = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: _UpperCamelCase = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): _UpperCamelCase = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: _UpperCamelCase = temp return maximum if __name__ == "__main__": print(solution())
63
1
def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Dict = [1] __snake_case : str = 0, 0, 0 __snake_case : List[str] = ugly_nums[ia] * 2 __snake_case : Optional[int] = ugly_nums[ia] * 3 __snake_case : Optional[int] = ugly_nums[ia] * 5 for _ in range(1 , _lowerCamelCase ): __snake_case : List[str] = min(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) ugly_nums.append(_lowerCamelCase ) if next_num == next_a: ia += 1 __snake_case : Tuple = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 __snake_case : Dict = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 __snake_case : Union[str, Any] = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(200) = }''')
123
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase__ : '''simple docstring''' def __init__( self :List[Any] , a :Dict , a :Any=3 , a :Any=3_2 , a :Optional[Any]=3 , a :str=1_0 , a :Union[str, Any]=[1_0, 2_0, 3_0, 4_0] , a :Optional[Any]=[1, 1, 2, 1] , a :Optional[Any]=True , a :Dict=True , a :Tuple="relu" , a :List[str]=3 , a :Tuple=None , ) -> Tuple: __UpperCamelCase : Optional[Any] = parent __UpperCamelCase : Dict = batch_size __UpperCamelCase : int = image_size __UpperCamelCase : Dict = num_channels __UpperCamelCase : Optional[int] = embeddings_size __UpperCamelCase : List[Any] = hidden_sizes __UpperCamelCase : Optional[Any] = depths __UpperCamelCase : Optional[int] = is_training __UpperCamelCase : Union[str, Any] = use_labels __UpperCamelCase : Optional[int] = hidden_act __UpperCamelCase : Tuple = num_labels __UpperCamelCase : Tuple = scope __UpperCamelCase : Dict = len(a ) def _lowerCamelCase ( self :Optional[int] ) -> Any: __UpperCamelCase : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCamelCase : List[str] = None if self.use_labels: __UpperCamelCase : Tuple = ids_tensor([self.batch_size] , self.num_labels ) __UpperCamelCase : List[Any] = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self :Union[str, Any] ) -> int: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def _lowerCamelCase ( self :List[Any] , a :Dict , a :int , a :Optional[Any] ) -> Tuple: __UpperCamelCase : str = TFResNetModel(config=a ) __UpperCamelCase : Union[str, Any] = model(a ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def _lowerCamelCase ( self :Union[str, Any] , a :Optional[int] , a :List[str] , a :Optional[Any] ) -> Any: __UpperCamelCase : str = self.num_labels __UpperCamelCase : Optional[int] = TFResNetForImageClassification(a ) __UpperCamelCase : List[str] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self :Optional[int] ) -> List[str]: __UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : Union[str, Any] = config_and_inputs __UpperCamelCase : List[str] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class lowerCamelCase__ ( __lowercase , __lowercase , unittest.TestCase): '''simple docstring''' _A = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () _A = ( {'feature-extraction': TFResNetModel, 'image-classification': TFResNetForImageClassification} if is_tf_available() else {} ) _A = False _A = False _A = False _A = False _A = False def _lowerCamelCase ( self :int ) -> List[str]: __UpperCamelCase : Union[str, Any] = TFResNetModelTester(self ) __UpperCamelCase : List[Any] = ConfigTester(self , config_class=a , has_text_modality=a ) def _lowerCamelCase ( self :int ) -> Dict: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self :str ) -> Optional[Any]: return @unittest.skip(reason="ResNet does not use inputs_embeds" ) def _lowerCamelCase ( self :Tuple ) -> Tuple: pass @unittest.skip(reason="ResNet does not support input and output embeddings" ) def _lowerCamelCase ( self :List[Any] ) -> List[str]: pass def _lowerCamelCase ( self :Optional[int] ) -> Tuple: __UpperCamelCase , __UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase : Dict = model_class(a ) __UpperCamelCase : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCamelCase : Dict = [*signature.parameters.keys()] __UpperCamelCase : Union[str, Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self :List[str] ) -> List[str]: __UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self :Optional[Any] ) -> Tuple: def check_hidden_states_output(a :Optional[Any] , a :Optional[int] , a :List[str] ): __UpperCamelCase : int = model_class(a ) __UpperCamelCase : int = model(**self._prepare_for_class(a , a ) ) __UpperCamelCase : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCamelCase : Optional[int] = self.model_tester.num_stages self.assertEqual(len(a ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __UpperCamelCase , __UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase : str = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: __UpperCamelCase : int = layer_type __UpperCamelCase : int = True check_hidden_states_output(a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCamelCase : int = True check_hidden_states_output(a , a , a ) def _lowerCamelCase ( self :Union[str, Any] ) -> Dict: __UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self :Dict ) -> Dict: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Optional[Any] = TFResNetModel.from_pretrained(a ) self.assertIsNotNone(a ) def _SCREAMING_SNAKE_CASE ( ) -> int: '''simple docstring''' __UpperCamelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") return image @require_tf @require_vision class lowerCamelCase__ ( unittest.TestCase): '''simple docstring''' @cached_property def _lowerCamelCase ( self :Optional[Any] ) -> Tuple: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self :Optional[int] ) -> Optional[int]: __UpperCamelCase : int = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __UpperCamelCase : List[Any] = self.default_image_processor __UpperCamelCase : List[str] = prepare_img() __UpperCamelCase : List[str] = image_processor(images=a , return_tensors="tf" ) # forward pass __UpperCamelCase : Dict = model(**a ) # verify the logits __UpperCamelCase : Dict = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , a ) __UpperCamelCase : Union[str, Any] = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , a , atol=1E-4 ) )
232
0
def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A , A : Optional[Any] = len(_lowerCamelCase ), len(grid[0] ) if ( min(_lowerCamelCase , _lowerCamelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) A : Tuple = 0 count += depth_first_search(_lowerCamelCase , row + 1 , _lowerCamelCase , _lowerCamelCase ) count += depth_first_search(_lowerCamelCase , row - 1 , _lowerCamelCase , _lowerCamelCase ) count += depth_first_search(_lowerCamelCase , _lowerCamelCase , col + 1 , _lowerCamelCase ) count += depth_first_search(_lowerCamelCase , _lowerCamelCase , col - 1 , _lowerCamelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
256
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def UpperCAmelCase ( _lowerCamelCase ): A : Any = [] for line in lines: A : List[str] = re.sub(R"#.*" , "" , _lowerCamelCase ) # remove comments if line: filtered_lines.append(_lowerCamelCase ) A : str = "\n".join(_lowerCamelCase ) # Make a hash from all this code A : Any = full_str.encode("utf-8" ) return shaaaa(_lowerCamelCase ).hexdigest() # get importable module names and hash for caching __SCREAMING_SNAKE_CASE = { """csv""": (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), """json""": (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), """pandas""": (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), """parquet""": (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), """arrow""": (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), """text""": (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), """imagefolder""": (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), """audiofolder""": (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions __SCREAMING_SNAKE_CASE = { """.csv""": ("""csv""", {}), """.tsv""": ("""csv""", {"""sep""": """\t"""}), """.json""": ("""json""", {}), """.jsonl""": ("""json""", {}), """.parquet""": ("""parquet""", {}), """.arrow""": ("""arrow""", {}), """.txt""": ("""text""", {}), } _EXTENSION_TO_MODULE.update({ext: ("""imagefolder""", {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ("""imagefolder""", {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ("""audiofolder""", {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ("""audiofolder""", {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) __SCREAMING_SNAKE_CASE = {"""imagefolder""", """audiofolder"""} # Used to filter data files based on extensions given a module name __SCREAMING_SNAKE_CASE = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append(""".zip""") _MODULE_TO_EXTENSIONS["audiofolder"].append(""".zip""")
256
1
def a ( snake_case__: int ): '''simple docstring''' if upper_limit < 0: raise ValueError('''Limit for the Catalan sequence must be ≥ 0''' ) lowercase_ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 lowercase_ = 1 if upper_limit > 0: lowercase_ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(snake_case__ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('\n********* Catalan Numbers Using Dynamic Programming ************\n') print('\n*** Enter -1 at any time to quit ***') print('\nEnter the upper limit (≥ 0) for the Catalan number sequence: ', end='') try: while True: __a = int(input().strip()) if N < 0: print('\n********* Goodbye!! ************') break else: print(f"The Catalan numbers from 0 through {N} are:") print(catalan_numbers(N)) print('Try another upper limit for the sequence: ', end='') except (NameError, ValueError): print('\n********* Invalid input, goodbye! ************\n') import doctest doctest.testmod()
30
import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def __UpperCamelCase ( _A , _A ): lowerCAmelCase_ = args.log_outputs lowerCAmelCase_ = '''_'''.join(args.dataset.split('''/''' ) + [args.config, args.split] ) # load metric lowerCAmelCase_ = load_metric('''wer''' ) lowerCAmelCase_ = load_metric('''cer''' ) # compute metrics lowerCAmelCase_ = wer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) lowerCAmelCase_ = cer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) # print & log results lowerCAmelCase_ = f"WER: {wer_result}\nCER: {cer_result}" print(_A ) with open(f"{dataset_id}_eval_results.txt" , '''w''' ) as f: f.write(_A ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: lowerCAmelCase_ = f"log_{dataset_id}_predictions.txt" lowerCAmelCase_ = f"log_{dataset_id}_targets.txt" with open(_A , '''w''' ) as p, open(_A , '''w''' ) as t: # mapping function to write output def write_to_file(_A , _A ): p.write(f"{i}" + '''\n''' ) p.write(batch['''prediction'''] + '''\n''' ) t.write(f"{i}" + '''\n''' ) t.write(batch['''target'''] + '''\n''' ) result.map(_A , with_indices=_A ) def __UpperCamelCase ( _A ): lowerCAmelCase_ = '''[,?.!\-\;\:"“%‘”�—’…–]''' # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training lowerCAmelCase_ = re.sub(_A , '''''' , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! lowerCAmelCase_ = ['''\n\n''', '''\n''', ''' ''', ''' '''] for t in token_sequences_to_ignore: lowerCAmelCase_ = ''' '''.join(text.split(_A ) ) return text def __UpperCamelCase ( _A ): # load dataset lowerCAmelCase_ = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=_A ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor lowerCAmelCase_ = AutoFeatureExtractor.from_pretrained(args.model_id ) lowerCAmelCase_ = feature_extractor.sampling_rate # resample audio lowerCAmelCase_ = dataset.cast_column('''audio''' , Audio(sampling_rate=_A ) ) # load eval pipeline if args.device is None: lowerCAmelCase_ = 0 if torch.cuda.is_available() else -1 lowerCAmelCase_ = pipeline('''automatic-speech-recognition''' , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(_A ): lowerCAmelCase_ = asr( batch['''audio''']['''array'''] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) lowerCAmelCase_ = prediction['''text'''] lowerCAmelCase_ = normalize_text(batch['''sentence'''] ) return batch # run inference on all examples lowerCAmelCase_ = dataset.map(_A , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(_A , _A ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument( '''--model_id''', type=str, required=True, help='''Model identifier. Should be loadable with 🤗 Transformers''' ) parser.add_argument( '''--dataset''', type=str, required=True, help='''Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets''', ) parser.add_argument( '''--config''', type=str, required=True, help='''Config of the dataset. *E.g.* `\'en\'` for Common Voice''' ) parser.add_argument('''--split''', type=str, required=True, help='''Split of the dataset. *E.g.* `\'test\'`''') parser.add_argument( '''--chunk_length_s''', type=float, default=None, help='''Chunk length in seconds. Defaults to 5 seconds.''' ) parser.add_argument( '''--stride_length_s''', type=float, default=None, help='''Stride of the audio chunks. Defaults to 1 second.''' ) parser.add_argument( '''--log_outputs''', action='''store_true''', help='''If defined, write outputs to log file for analysis.''' ) parser.add_argument( '''--device''', type=int, default=None, help='''The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.''', ) _A = parser.parse_args() main(args)
278
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowercase ( __UpperCAmelCase ): lowercase_ = 'roformer' def __init__( self , UpperCAmelCase_=50000 , UpperCAmelCase_=None , UpperCAmelCase_=768 , UpperCAmelCase_=12 , UpperCAmelCase_=12 , UpperCAmelCase_=3072 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=1536 , UpperCAmelCase_=2 , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-1_2 , UpperCAmelCase_=0 , UpperCAmelCase_=False , UpperCAmelCase_=True , **UpperCAmelCase_ , ) -> int: super().__init__(pad_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) lowerCamelCase : Optional[Any] = vocab_size lowerCamelCase : List[str] = hidden_size if embedding_size is None else embedding_size lowerCamelCase : Dict = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : Optional[int] = num_attention_heads lowerCamelCase : Optional[Any] = hidden_act lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : Optional[int] = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Any = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[int] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = rotary_value lowerCamelCase : Optional[Any] = use_cache class _lowercase ( __UpperCAmelCase ): @property def _UpperCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCamelCase : Tuple = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowerCamelCase : str = {0: 'batch', 1: 'sequence'} lowerCamelCase : Tuple = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
205
"""simple docstring""" import numpy as np def UpperCAmelCase ( a_, a_, a_ = 1E-12, a_ = 100, ): '''simple docstring''' assert np.shape(a_ )[0] == np.shape(a_ )[1] # Ensure proper dimensionality. assert np.shape(a_ )[0] == np.shape(a_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(a_ ) == np.iscomplexobj(a_ ) lowerCamelCase : Optional[int] = np.iscomplexobj(a_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(a_, input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. lowerCamelCase : Union[str, Any] = False lowerCamelCase : List[str] = 0 lowerCamelCase : Any = 0 lowerCamelCase : Dict = 1E12 while not convergence: # Multiple matrix by the vector. lowerCamelCase : Optional[int] = np.dot(a_, a_ ) # Normalize the resulting output vector. lowerCamelCase : Optional[int] = w / np.linalg.norm(a_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) lowerCamelCase : Optional[Any] = vector.conj().T if is_complex else vector.T lowerCamelCase : str = np.dot(a_, np.dot(a_, a_ ) ) # Check convergence. lowerCamelCase : Optional[int] = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: lowerCamelCase : int = True lowerCamelCase : Optional[Any] = lambda_ if is_complex: lowerCamelCase : Any = np.real(lambda_ ) return lambda_, vector def UpperCAmelCase ( ): '''simple docstring''' lowerCamelCase : str = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) lowerCamelCase : str = np.array([41, 4, 20] ) lowerCamelCase : Optional[Any] = real_input_matrix.astype(np.complexaaa ) lowerCamelCase : Dict = np.triu(1j * complex_input_matrix, 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T lowerCamelCase : List[Any] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": lowerCamelCase : str = real_input_matrix lowerCamelCase : Any = real_vector elif problem_type == "complex": lowerCamelCase : str = complex_input_matrix lowerCamelCase : Dict = complex_vector # Our implementation. lowerCamelCase , lowerCamelCase : List[str] = power_iteration(a_, a_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). lowerCamelCase , lowerCamelCase : Optional[Any] = np.linalg.eigh(a_ ) # Last eigenvalue is the maximum one. lowerCamelCase : Dict = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. lowerCamelCase : List[str] = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1E-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(a_ ) - np.abs(a_ ) ) <= 1E-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
205
1
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class a__ : """simple docstring""" __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None # sigma(t_i) @classmethod def UpperCamelCase ( cls ) -> Any: '''simple docstring''' return cls() @dataclass class a__ ( snake_case ): """simple docstring""" __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 class a__ ( snake_case , snake_case ): """simple docstring""" @property def UpperCamelCase ( self ) -> Tuple: '''simple docstring''' return True @register_to_config def __init__( self , lowercase = 0.02 , lowercase = 100 , lowercase = 1.007 , lowercase = 80 , lowercase = 0.05 , lowercase = 50 , ) -> Any: '''simple docstring''' pass def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' return KarrasVeSchedulerState.create() def UpperCamelCase ( self , lowercase , lowercase , lowercase = () ) -> KarrasVeSchedulerState: '''simple docstring''' A__ = jnp.arange(0 , lowercase )[::-1].copy() A__ = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=lowercase , schedule=jnp.array(lowercase , dtype=jnp.floataa ) , timesteps=lowercase , ) def UpperCamelCase ( self , lowercase , lowercase , lowercase , lowercase , ) -> Tuple[jnp.ndarray, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: A__ = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: A__ = 0 # sample eps ~ N(0, S_noise^2 * I) A__ = random.split(lowercase , num=1 ) A__ = self.config.s_noise * random.normal(key=lowercase , shape=sample.shape ) A__ = sigma + gamma * sigma A__ = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: '''simple docstring''' A__ = sample_hat + sigma_hat * model_output A__ = (sample_hat - pred_original_sample) / sigma_hat A__ = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=lowercase , derivative=lowercase , state=lowercase ) def UpperCamelCase ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: '''simple docstring''' A__ = sample_prev + sigma_prev * model_output A__ = (sample_prev - pred_original_sample) / sigma_prev A__ = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=lowercase , derivative=lowercase , state=lowercase ) def UpperCamelCase ( self , lowercase , lowercase , lowercase , lowercase ) -> Tuple: '''simple docstring''' raise NotImplementedError()
68
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup __snake_case = '''https://www.indeed.co.in/jobs?q=mobile+app+development&l=''' def _A ( _lowercase = "mumbai" ) -> Generator[tuple[str, str], None, None]: """simple docstring""" __UpperCamelCase = BeautifulSoup(requests.get(url + location ).content , 'html.parser' ) # This attribute finds out all the specifics listed in a job for job in soup.find_all('div' , attrs={'data-tn-component': 'organicJob'} ): __UpperCamelCase = job.find('a' , attrs={'data-tn-element': 'jobTitle'} ).text.strip() __UpperCamelCase = job.find('span' , {'class': 'company'} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs('''Bangalore'''), 1): print(f"""Job {i:>2} is {job[0]} at {job[1]}""")
310
0
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Union[str, Any] = abs(UpperCamelCase__ ) _a : Tuple = 0 while n > 0: res += n % 1_0 n //= 1_0 return res def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Optional[int] = abs(UpperCamelCase__ ) return n if n < 1_0 else n % 1_0 + sum_of_digits(n // 1_0 ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return sum(int(UpperCamelCase__ ) for c in str(abs(UpperCamelCase__ ) ) ) def lowerCAmelCase__ ( ): '''simple docstring''' from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCamelCase__ , UpperCamelCase__ ) -> None: _a : Optional[int] = F"""{func.__name__}({value})""" _a : int = timeit(F"""__main__.{call}""" , setup="""import __main__""" ) print(F"""{call:56} = {func(UpperCamelCase__ )} -- {timing:.4f} seconds""" ) for value in (2_6_2_1_4_4, 1_1_2_5_8_9_9_9_0_6_8_4_2_6_2_4, 1_2_6_7_6_5_0_6_0_0_2_2_8_2_2_9_4_0_1_4_9_6_7_0_3_2_0_5_3_7_6): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(UpperCamelCase__ , UpperCamelCase__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
324
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _snake_case = { 'configuration_vision_encoder_decoder': ['VisionEncoderDecoderConfig', 'VisionEncoderDecoderOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['VisionEncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['TFVisionEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['FlaxVisionEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
324
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A : Dict = { 'configuration_trocr': ['TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrOCRConfig'], 'processing_trocr': ['TrOCRProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Any = [ 'TROCR_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrOCRForCausalLM', 'TrOCRPreTrainedModel', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
6
'''simple docstring''' def snake_case ( UpperCAmelCase )-> list[int]: """simple docstring""" if length <= 0 or not isinstance(UpperCAmelCase , UpperCAmelCase ): raise ValueError('Length must be a positive integer.' ) return [n * (2 * n - 1) for n in range(UpperCAmelCase )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=1_0))
161
0
"""simple docstring""" import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import require_torch if is_torch_available(): from transformers.trainer import DEFAULT_CALLBACKS from .test_trainer import RegressionDataset, RegressionModelConfig, RegressionPreTrainedModel class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' def __init__( self ) -> Optional[Any]: _lowerCAmelCase =[] def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: self.events.append("""on_init_end""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Any: self.events.append("""on_train_begin""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Any: self.events.append("""on_train_end""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> str: self.events.append("""on_epoch_begin""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> int: self.events.append("""on_epoch_end""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Dict: self.events.append("""on_step_begin""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Dict: self.events.append("""on_step_end""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: self.events.append("""on_evaluate""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: self.events.append("""on_predict""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: self.events.append("""on_save""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> Any: self.events.append("""on_log""" ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: self.events.append("""on_prediction_step""" ) @require_torch class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ) -> List[str]: _lowerCAmelCase =tempfile.mkdtemp() def _lowerCAmelCase ( self ) -> Any: shutil.rmtree(self.output_dir ) def _lowerCAmelCase ( self , __UpperCAmelCase=0 , __UpperCAmelCase=0 , __UpperCAmelCase=64 , __UpperCAmelCase=64 , __UpperCAmelCase=None , __UpperCAmelCase=False , **__UpperCAmelCase ) -> List[Any]: # disable_tqdm in TrainingArguments has a flaky default since it depends on the level of logging. We make sure # its set to False since the tests later on depend on its value. _lowerCAmelCase =RegressionDataset(length=__UpperCAmelCase ) _lowerCAmelCase =RegressionDataset(length=__UpperCAmelCase ) _lowerCAmelCase =RegressionModelConfig(a=__UpperCAmelCase , b=__UpperCAmelCase ) _lowerCAmelCase =RegressionPreTrainedModel(__UpperCAmelCase ) _lowerCAmelCase =TrainingArguments(self.output_dir , disable_tqdm=__UpperCAmelCase , report_to=[] , **__UpperCAmelCase ) return Trainer( __UpperCAmelCase , __UpperCAmelCase , train_dataset=__UpperCAmelCase , eval_dataset=__UpperCAmelCase , callbacks=__UpperCAmelCase , ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Any: self.assertEqual(len(__UpperCAmelCase ) , len(__UpperCAmelCase ) ) # Order doesn't matter _lowerCAmelCase =sorted(__UpperCAmelCase , key=lambda __UpperCAmelCase : cb.__name__ if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else cb.__class__.__name__ ) _lowerCAmelCase =sorted(__UpperCAmelCase , key=lambda __UpperCAmelCase : cb.__name__ if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else cb.__class__.__name__ ) for cba, cba in zip(__UpperCAmelCase , __UpperCAmelCase ): if isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(__UpperCAmelCase , __UpperCAmelCase ): self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ) and not isinstance(__UpperCAmelCase , __UpperCAmelCase ): self.assertEqual(__UpperCAmelCase , cba.__class__ ) elif not isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(__UpperCAmelCase , __UpperCAmelCase ): self.assertEqual(cba.__class__ , __UpperCAmelCase ) else: self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) def _lowerCAmelCase ( self , __UpperCAmelCase ) -> Optional[int]: _lowerCAmelCase =["""on_init_end""", """on_train_begin"""] _lowerCAmelCase =0 _lowerCAmelCase =len(trainer.get_eval_dataloader() ) _lowerCAmelCase =["""on_prediction_step"""] * len(trainer.get_eval_dataloader() ) + ["""on_log""", """on_evaluate"""] for _ in range(trainer.state.num_train_epochs ): expected_events.append("""on_epoch_begin""" ) for _ in range(__UpperCAmelCase ): step += 1 expected_events += ["on_step_begin", "on_step_end"] if step % trainer.args.logging_steps == 0: expected_events.append("""on_log""" ) if trainer.args.evaluation_strategy == IntervalStrategy.STEPS and step % trainer.args.eval_steps == 0: expected_events += evaluation_events.copy() if step % trainer.args.save_steps == 0: expected_events.append("""on_save""" ) expected_events.append("""on_epoch_end""" ) if trainer.args.evaluation_strategy == IntervalStrategy.EPOCH: expected_events += evaluation_events.copy() expected_events += ["on_log", "on_train_end"] return expected_events def _lowerCAmelCase ( self ) -> Union[str, Any]: _lowerCAmelCase =self.get_trainer() _lowerCAmelCase =DEFAULT_CALLBACKS.copy() + [ProgressCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) # Callbacks passed at init are added to the default callbacks _lowerCAmelCase =self.get_trainer(callbacks=[MyTestTrainerCallback] ) expected_callbacks.append(__UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) # TrainingArguments.disable_tqdm controls if use ProgressCallback or PrinterCallback _lowerCAmelCase =self.get_trainer(disable_tqdm=__UpperCAmelCase ) _lowerCAmelCase =DEFAULT_CALLBACKS.copy() + [PrinterCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) def _lowerCAmelCase ( self ) -> Union[str, Any]: _lowerCAmelCase =DEFAULT_CALLBACKS.copy() + [ProgressCallback] _lowerCAmelCase =self.get_trainer() # We can add, pop, or remove by class name trainer.remove_callback(__UpperCAmelCase ) expected_callbacks.remove(__UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) _lowerCAmelCase =self.get_trainer() _lowerCAmelCase =trainer.pop_callback(__UpperCAmelCase ) self.assertEqual(cb.__class__ , __UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) trainer.add_callback(__UpperCAmelCase ) expected_callbacks.insert(0 , __UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) # We can also add, pop, or remove by instance _lowerCAmelCase =self.get_trainer() _lowerCAmelCase =trainer.callback_handler.callbacks[0] trainer.remove_callback(__UpperCAmelCase ) expected_callbacks.remove(__UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) _lowerCAmelCase =self.get_trainer() _lowerCAmelCase =trainer.callback_handler.callbacks[0] _lowerCAmelCase =trainer.pop_callback(__UpperCAmelCase ) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) trainer.add_callback(__UpperCAmelCase ) expected_callbacks.insert(0 , __UpperCAmelCase ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __UpperCAmelCase ) def _lowerCAmelCase ( self ) -> Optional[Any]: import warnings # XXX: for now ignore scatter_gather warnings in this test since it's not relevant to what's being tested warnings.simplefilter(action="""ignore""" , category=__UpperCAmelCase ) _lowerCAmelCase =self.get_trainer(callbacks=[MyTestTrainerCallback] ) trainer.train() _lowerCAmelCase =trainer.callback_handler.callbacks[-2].events self.assertEqual(__UpperCAmelCase , self.get_expected_events(__UpperCAmelCase ) ) # Independent log/save/eval _lowerCAmelCase =self.get_trainer(callbacks=[MyTestTrainerCallback] , logging_steps=5 ) trainer.train() _lowerCAmelCase =trainer.callback_handler.callbacks[-2].events self.assertEqual(__UpperCAmelCase , self.get_expected_events(__UpperCAmelCase ) ) _lowerCAmelCase =self.get_trainer(callbacks=[MyTestTrainerCallback] , save_steps=5 ) trainer.train() _lowerCAmelCase =trainer.callback_handler.callbacks[-2].events self.assertEqual(__UpperCAmelCase , self.get_expected_events(__UpperCAmelCase ) ) _lowerCAmelCase =self.get_trainer(callbacks=[MyTestTrainerCallback] , eval_steps=5 , evaluation_strategy="""steps""" ) trainer.train() _lowerCAmelCase =trainer.callback_handler.callbacks[-2].events self.assertEqual(__UpperCAmelCase , self.get_expected_events(__UpperCAmelCase ) ) _lowerCAmelCase =self.get_trainer(callbacks=[MyTestTrainerCallback] , evaluation_strategy="""epoch""" ) trainer.train() _lowerCAmelCase =trainer.callback_handler.callbacks[-2].events self.assertEqual(__UpperCAmelCase , self.get_expected_events(__UpperCAmelCase ) ) # A bit of everything _lowerCAmelCase =self.get_trainer( callbacks=[MyTestTrainerCallback] , logging_steps=3 , save_steps=10 , eval_steps=5 , evaluation_strategy="""steps""" , ) trainer.train() _lowerCAmelCase =trainer.callback_handler.callbacks[-2].events self.assertEqual(__UpperCAmelCase , self.get_expected_events(__UpperCAmelCase ) ) # warning should be emitted for duplicated callbacks with patch("""transformers.trainer_callback.logger.warning""" ) as warn_mock: _lowerCAmelCase =self.get_trainer( callbacks=[MyTestTrainerCallback, MyTestTrainerCallback] , ) assert str(__UpperCAmelCase ) in warn_mock.call_args[0][0]
341
"""simple docstring""" import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase=1 ) -> Tuple: if n_shave_prefix_segments >= 0: return ".".join(path.split(""".""" )[n_shave_prefix_segments:] ) else: return ".".join(path.split(""".""" )[:n_shave_prefix_segments] ) def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase=0 ) -> List[str]: _lowerCAmelCase =[] for old_item in old_list: _lowerCAmelCase =old_item.replace("""in_layers.0""" , """norm1""" ) _lowerCAmelCase =new_item.replace("""in_layers.2""" , """conv1""" ) _lowerCAmelCase =new_item.replace("""out_layers.0""" , """norm2""" ) _lowerCAmelCase =new_item.replace("""out_layers.3""" , """conv2""" ) _lowerCAmelCase =new_item.replace("""emb_layers.1""" , """time_emb_proj""" ) _lowerCAmelCase =new_item.replace("""skip_connection""" , """conv_shortcut""" ) _lowerCAmelCase =shave_segments(__UpperCamelCase , n_shave_prefix_segments=__UpperCamelCase ) mapping.append({"""old""": old_item, """new""": new_item} ) return mapping def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase=0 ) -> Tuple: _lowerCAmelCase =[] for old_item in old_list: _lowerCAmelCase =old_item _lowerCAmelCase =new_item.replace("""norm.weight""" , """group_norm.weight""" ) _lowerCAmelCase =new_item.replace("""norm.bias""" , """group_norm.bias""" ) _lowerCAmelCase =new_item.replace("""proj_out.weight""" , """proj_attn.weight""" ) _lowerCAmelCase =new_item.replace("""proj_out.bias""" , """proj_attn.bias""" ) _lowerCAmelCase =shave_segments(__UpperCamelCase , n_shave_prefix_segments=__UpperCamelCase ) mapping.append({"""old""": old_item, """new""": new_item} ) return mapping def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None ) -> Optional[int]: assert isinstance(__UpperCamelCase , __UpperCamelCase ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): _lowerCAmelCase =old_checkpoint[path] _lowerCAmelCase =old_tensor.shape[0] // 3 _lowerCAmelCase =(-1, channels) if len(old_tensor.shape ) == 3 else (-1) _lowerCAmelCase =old_tensor.shape[0] // config["""num_head_channels"""] // 3 _lowerCAmelCase =old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase =old_tensor.split(channels // num_heads , dim=1 ) _lowerCAmelCase =query.reshape(__UpperCamelCase ) _lowerCAmelCase =key.reshape(__UpperCamelCase ) _lowerCAmelCase =value.reshape(__UpperCamelCase ) for path in paths: _lowerCAmelCase =path["""new"""] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here _lowerCAmelCase =new_path.replace("""middle_block.0""" , """mid_block.resnets.0""" ) _lowerCAmelCase =new_path.replace("""middle_block.1""" , """mid_block.attentions.0""" ) _lowerCAmelCase =new_path.replace("""middle_block.2""" , """mid_block.resnets.1""" ) if additional_replacements is not None: for replacement in additional_replacements: _lowerCAmelCase =new_path.replace(replacement["""old"""] , replacement["""new"""] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: _lowerCAmelCase =old_checkpoint[path["""old"""]][:, :, 0] else: _lowerCAmelCase =old_checkpoint[path["""old"""]] def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase ) -> Optional[Any]: _lowerCAmelCase ={} _lowerCAmelCase =checkpoint["""time_embed.0.weight"""] _lowerCAmelCase =checkpoint["""time_embed.0.bias"""] _lowerCAmelCase =checkpoint["""time_embed.2.weight"""] _lowerCAmelCase =checkpoint["""time_embed.2.bias"""] _lowerCAmelCase =checkpoint["""input_blocks.0.0.weight"""] _lowerCAmelCase =checkpoint["""input_blocks.0.0.bias"""] _lowerCAmelCase =checkpoint["""out.0.weight"""] _lowerCAmelCase =checkpoint["""out.0.bias"""] _lowerCAmelCase =checkpoint["""out.2.weight"""] _lowerCAmelCase =checkpoint["""out.2.bias"""] # Retrieves the keys for the input blocks only _lowerCAmelCase =len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """input_blocks""" in layer} ) _lowerCAmelCase ={ layer_id: [key for key in checkpoint if F'''input_blocks.{layer_id}''' in key] for layer_id in range(__UpperCamelCase ) } # Retrieves the keys for the middle blocks only _lowerCAmelCase =len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """middle_block""" in layer} ) _lowerCAmelCase ={ layer_id: [key for key in checkpoint if F'''middle_block.{layer_id}''' in key] for layer_id in range(__UpperCamelCase ) } # Retrieves the keys for the output blocks only _lowerCAmelCase =len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """output_blocks""" in layer} ) _lowerCAmelCase ={ layer_id: [key for key in checkpoint if F'''output_blocks.{layer_id}''' in key] for layer_id in range(__UpperCamelCase ) } for i in range(1 , __UpperCamelCase ): _lowerCAmelCase =(i - 1) // (config["""num_res_blocks"""] + 1) _lowerCAmelCase =(i - 1) % (config["""num_res_blocks"""] + 1) _lowerCAmelCase =[key for key in input_blocks[i] if F'''input_blocks.{i}.0''' in key] _lowerCAmelCase =[key for key in input_blocks[i] if F'''input_blocks.{i}.1''' in key] if F'''input_blocks.{i}.0.op.weight''' in checkpoint: _lowerCAmelCase =checkpoint[ F'''input_blocks.{i}.0.op.weight''' ] _lowerCAmelCase =checkpoint[ F'''input_blocks.{i}.0.op.bias''' ] continue _lowerCAmelCase =renew_resnet_paths(__UpperCamelCase ) _lowerCAmelCase ={"""old""": F'''input_blocks.{i}.0''', """new""": F'''down_blocks.{block_id}.resnets.{layer_in_block_id}'''} _lowerCAmelCase ={"""old""": """resnets.2.op""", """new""": """downsamplers.0.op"""} assign_to_checkpoint( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , additional_replacements=[meta_path, resnet_op] , config=__UpperCamelCase ) if len(__UpperCamelCase ): _lowerCAmelCase =renew_attention_paths(__UpperCamelCase ) _lowerCAmelCase ={ """old""": F'''input_blocks.{i}.1''', """new""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}''', } _lowerCAmelCase ={ F'''input_blocks.{i}.1.qkv.bias''': { """key""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias''', """query""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias''', """value""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias''', }, F'''input_blocks.{i}.1.qkv.weight''': { """key""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight''', """query""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight''', """value""": F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight''', }, } assign_to_checkpoint( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , additional_replacements=[meta_path] , attention_paths_to_split=__UpperCamelCase , config=__UpperCamelCase , ) _lowerCAmelCase =middle_blocks[0] _lowerCAmelCase =middle_blocks[1] _lowerCAmelCase =middle_blocks[2] _lowerCAmelCase =renew_resnet_paths(__UpperCamelCase ) assign_to_checkpoint(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , config=__UpperCamelCase ) _lowerCAmelCase =renew_resnet_paths(__UpperCamelCase ) assign_to_checkpoint(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , config=__UpperCamelCase ) _lowerCAmelCase =renew_attention_paths(__UpperCamelCase ) _lowerCAmelCase ={ """middle_block.1.qkv.bias""": { """key""": """mid_block.attentions.0.key.bias""", """query""": """mid_block.attentions.0.query.bias""", """value""": """mid_block.attentions.0.value.bias""", }, """middle_block.1.qkv.weight""": { """key""": """mid_block.attentions.0.key.weight""", """query""": """mid_block.attentions.0.query.weight""", """value""": """mid_block.attentions.0.value.weight""", }, } assign_to_checkpoint( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , attention_paths_to_split=__UpperCamelCase , config=__UpperCamelCase ) for i in range(__UpperCamelCase ): _lowerCAmelCase =i // (config["""num_res_blocks"""] + 1) _lowerCAmelCase =i % (config["""num_res_blocks"""] + 1) _lowerCAmelCase =[shave_segments(__UpperCamelCase , 2 ) for name in output_blocks[i]] _lowerCAmelCase ={} for layer in output_block_layers: _lowerCAmelCase , _lowerCAmelCase =layer.split(""".""" )[0], shave_segments(__UpperCamelCase , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(__UpperCamelCase ) else: _lowerCAmelCase =[layer_name] if len(__UpperCamelCase ) > 1: _lowerCAmelCase =[key for key in output_blocks[i] if F'''output_blocks.{i}.0''' in key] _lowerCAmelCase =[key for key in output_blocks[i] if F'''output_blocks.{i}.1''' in key] _lowerCAmelCase =renew_resnet_paths(__UpperCamelCase ) _lowerCAmelCase =renew_resnet_paths(__UpperCamelCase ) _lowerCAmelCase ={"""old""": F'''output_blocks.{i}.0''', """new""": F'''up_blocks.{block_id}.resnets.{layer_in_block_id}'''} assign_to_checkpoint(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , additional_replacements=[meta_path] , config=__UpperCamelCase ) if ["conv.weight", "conv.bias"] in output_block_list.values(): _lowerCAmelCase =list(output_block_list.values() ).index(["""conv.weight""", """conv.bias"""] ) _lowerCAmelCase =checkpoint[ F'''output_blocks.{i}.{index}.conv.weight''' ] _lowerCAmelCase =checkpoint[ F'''output_blocks.{i}.{index}.conv.bias''' ] # Clear attentions as they have been attributed above. if len(__UpperCamelCase ) == 2: _lowerCAmelCase =[] if len(__UpperCamelCase ): _lowerCAmelCase =renew_attention_paths(__UpperCamelCase ) _lowerCAmelCase ={ """old""": F'''output_blocks.{i}.1''', """new""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}''', } _lowerCAmelCase ={ F'''output_blocks.{i}.1.qkv.bias''': { """key""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias''', """query""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias''', """value""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias''', }, F'''output_blocks.{i}.1.qkv.weight''': { """key""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight''', """query""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight''', """value""": F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight''', }, } assign_to_checkpoint( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any("""qkv""" in key for key in attentions ) else None , config=__UpperCamelCase , ) else: _lowerCAmelCase =renew_resnet_paths(__UpperCamelCase , n_shave_prefix_segments=1 ) for path in resnet_0_paths: _lowerCAmelCase =""".""".join(["""output_blocks""", str(__UpperCamelCase ), path["""old"""]] ) _lowerCAmelCase =""".""".join(["""up_blocks""", str(__UpperCamelCase ), """resnets""", str(__UpperCamelCase ), path["""new"""]] ) _lowerCAmelCase =checkpoint[old_path] return new_checkpoint if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the architecture.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') __A = parser.parse_args() __A = torch.load(args.checkpoint_path) with open(args.config_file) as f: __A = json.loads(f.read()) __A = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] __A = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: __A = DDPMScheduler.from_config('/'.join(args.checkpoint_path.split('/')[:-1])) __A = VQModel.from_pretrained('/'.join(args.checkpoint_path.split('/')[:-1])) __A = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
341
1
import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): warnings.warn( """The class DonutFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use DonutImageProcessor instead.""" , __magic_name__ , ) super().__init__(*__magic_name__ , **__magic_name__ )
287
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
1
'''simple docstring''' from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class __UpperCAmelCase ( A__ ): '''simple docstring''' __lowerCAmelCase = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
360
'''simple docstring''' from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def __a ( UpperCAmelCase ) ->List[str]: """simple docstring""" if isinstance(UpperCAmelCase , collections.abc.Iterable ): return x return (x, x) @require_tf class __UpperCAmelCase : '''simple docstring''' def A (self : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] ): pass def A (self : List[str] ): pass def A (self : Union[str, Any] ): pass def A (self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int=None , **_lowerCAmelCase : Dict ): A = VisionTextDualEncoderConfig.from_vision_text_configs(_lowerCAmelCase , _lowerCAmelCase ) A = TFVisionTextDualEncoderModel(_lowerCAmelCase ) A = model(input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], config.projection_dim) ) def A (self : Dict , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict=None , **_lowerCAmelCase : int ): A , A = self.get_vision_text_model(_lowerCAmelCase , _lowerCAmelCase ) A = TFVisionTextDualEncoderModel(vision_model=_lowerCAmelCase , text_model=_lowerCAmelCase ) A = model(input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], model.config.projection_dim) ) def A (self : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None , **_lowerCAmelCase : List[Any] ): A , A = self.get_vision_text_model(_lowerCAmelCase , _lowerCAmelCase ) A = {"""vision_model""": vision_model, """text_model""": text_model} A = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**_lowerCAmelCase ) A = model(input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], model.config.projection_dim) ) def A (self : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any]=None , **_lowerCAmelCase : Any ): A , A = self.get_vision_text_model(_lowerCAmelCase , _lowerCAmelCase ) A = TFVisionTextDualEncoderModel(vision_model=_lowerCAmelCase , text_model=_lowerCAmelCase ) A = model(input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase ) A = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_lowerCAmelCase ) A = TFVisionTextDualEncoderModel.from_pretrained(_lowerCAmelCase ) A = model(input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase ) A = after_output[0].numpy() A = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_lowerCAmelCase , 1e-5 ) def A (self : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Any=None , **_lowerCAmelCase : List[Any] ): A , A = self.get_vision_text_model(_lowerCAmelCase , _lowerCAmelCase ) A = TFVisionTextDualEncoderModel(vision_model=_lowerCAmelCase , text_model=_lowerCAmelCase ) A = model( input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , output_attentions=_lowerCAmelCase ) A = output.vision_model_output.attentions self.assertEqual(len(_lowerCAmelCase ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) A = to_atuple(vision_model.config.image_size ) A = to_atuple(vision_model.config.patch_size ) A = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) A = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) A = output.text_model_output.attentions self.assertEqual(len(_lowerCAmelCase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def A (self : List[Any] , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : np.ndarray , _lowerCAmelCase : float ): A = np.abs((a - b) ).max() self.assertLessEqual(_lowerCAmelCase , _lowerCAmelCase , F"""Difference between torch and flax is {diff} (>= {tol}).""" ) def A (self : List[str] ): A = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**_lowerCAmelCase ) def A (self : Optional[int] ): A = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_lowerCAmelCase ) def A (self : List[Any] ): A = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_lowerCAmelCase ) def A (self : int ): A = self.prepare_config_and_inputs() self.check_save_load(**_lowerCAmelCase ) def A (self : int ): A = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_lowerCAmelCase ) @slow def A (self : Tuple ): A , A = self.get_pretrained_model_and_inputs() A = model_a(**_lowerCAmelCase ) A = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_lowerCAmelCase ) A = TFVisionTextDualEncoderModel.from_pretrained(_lowerCAmelCase ) A = model_a(**_lowerCAmelCase ) A = after_outputs[0].numpy() A = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_lowerCAmelCase , 1e-5 ) @require_tf class __UpperCAmelCase ( A__ , unittest.TestCase ): '''simple docstring''' def A (self : int ): A = TFVisionTextDualEncoderModel.from_vision_text_pretrained( """hf-internal-testing/tiny-random-vit""" , """hf-internal-testing/tiny-random-bert""" ) A = 13 A = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) A = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) A = random_attention_mask([batch_size, 4] ) A = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def A (self : Dict , _lowerCAmelCase : Dict , _lowerCAmelCase : int ): A = TFViTModel(_lowerCAmelCase , name="""vision_model""" ) A = TFBertModel(_lowerCAmelCase , name="""text_model""" ) return vision_model, text_model def A (self : Union[str, Any] ): A = TFViTModelTester(self ) A = TFBertModelTester(self ) A = vit_model_tester.prepare_config_and_inputs() A = bert_model_tester.prepare_config_and_inputs() A , A , A = vision_config_and_inputs ( ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class __UpperCAmelCase ( A__ , unittest.TestCase ): '''simple docstring''' def A (self : Optional[int] ): # DeiT repo doesn't have TF weights, but we don't actually use the weights at all so let's # just reinitialize it. A = TFVisionTextDualEncoderModel.from_vision_text_pretrained( """Rocketknight1/tiny-random-deit-tf""" , """hf-internal-testing/tiny-random-roberta""" ) A = 13 A = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) A = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) A = random_attention_mask([batch_size, 4] ) A = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def A (self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any]=None , **_lowerCAmelCase : Any ): A , A = self.get_vision_text_model(_lowerCAmelCase , _lowerCAmelCase ) A = TFVisionTextDualEncoderModel(vision_model=_lowerCAmelCase , text_model=_lowerCAmelCase ) A = model( input_ids=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , output_attentions=_lowerCAmelCase ) A = output.vision_model_output.attentions self.assertEqual(len(_lowerCAmelCase ) , vision_config.num_hidden_layers ) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) A = to_atuple(vision_model.config.image_size ) A = to_atuple(vision_model.config.patch_size ) A = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) A = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) A = output.text_model_output.attentions self.assertEqual(len(_lowerCAmelCase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def A (self : Any , _lowerCAmelCase : Any , _lowerCAmelCase : str ): A = TFDeiTModel(_lowerCAmelCase , name="""vision_model""" ) A = TFRobertaModel(_lowerCAmelCase , name="""text_model""" ) return vision_model, text_model def A (self : str ): A = TFDeiTModelTester(self ) A = TFRobertaModelTester(self ) A = vit_model_tester.prepare_config_and_inputs() A = bert_model_tester.prepare_config_and_inputs() A , A , A = vision_config_and_inputs ( ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class __UpperCAmelCase ( A__ , unittest.TestCase ): '''simple docstring''' def A (self : Dict ): A = TFVisionTextDualEncoderModel.from_vision_text_pretrained( """Rocketknight1/tiny-random-clip-tf""" , """hf-internal-testing/tiny-random-bert""" ) A = 13 A = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) A = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) A = random_attention_mask([batch_size, 4] ) A = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def A (self : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ): A = TFCLIPVisionModel(_lowerCAmelCase , name="""vision_model""" ) A = TFBertModel(_lowerCAmelCase , name="""text_model""" ) return vision_model, text_model def A (self : Optional[Any] ): A = TFCLIPVisionModelTester(self ) A = TFBertModelTester(self ) A = clip_model_tester.prepare_config_and_inputs() A = bert_model_tester.prepare_config_and_inputs() A , A = vision_config_and_inputs ( ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def A (self : Any ): A = TFVisionTextDualEncoderModel.from_pretrained( """clip-italian/clip-italian""" , logit_scale_init_value=1.0 , from_pt=_lowerCAmelCase ) A = VisionTextDualEncoderProcessor.from_pretrained("""clip-italian/clip-italian""" ) A = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) A = processor( text=["""una foto di un gatto""", """una foto di un cane"""] , images=_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors="""np""" ) A = model(**_lowerCAmelCase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) A = np.array([[1.2_284_727, 0.3_104_122]] ) self.assertTrue(np.allclose(outputs.logits_per_image.numpy() , _lowerCAmelCase , atol=1e-3 ) )
337
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import copy import importlib.metadata import json import os from dataclasses import dataclass from typing import Any, Dict, Union from packaging import version from ..utils import is_torch_available, logging if is_torch_available(): import torch lowercase = logging.get_logger(__name__) @dataclass class UpperCamelCase_ : '''simple docstring''' def __init__( self , a=False , a=False , a=6.0 , a=None , a=False , a=False , a=None , a="fp4" , a=False , **a , ) -> List[Any]: snake_case_ = load_in_abit snake_case_ = load_in_abit snake_case_ = llm_inta_threshold snake_case_ = llm_inta_skip_modules snake_case_ = llm_inta_enable_fpaa_cpu_offload snake_case_ = llm_inta_has_fpaa_weight snake_case_ = bnb_abit_quant_type snake_case_ = bnb_abit_use_double_quant if bnb_abit_compute_dtype is None: snake_case_ = torch.floataa elif isinstance(a , a ): snake_case_ = getattr(a , a ) elif isinstance(a , torch.dtype ): snake_case_ = bnb_abit_compute_dtype else: raise ValueError('bnb_4bit_compute_dtype must be a string or a torch.dtype' ) self.post_init() def _UpperCamelCase ( self ) -> Tuple: if not isinstance(self.llm_inta_threshold , a ): raise ValueError('llm_int8_threshold must be a float' ) if self.llm_inta_skip_modules is not None and not isinstance(self.llm_inta_skip_modules , a ): raise ValueError('llm_int8_skip_modules must be a list of strings' ) if not isinstance(self.llm_inta_enable_fpaa_cpu_offload , a ): raise ValueError('llm_int8_enable_fp32_cpu_offload must be a boolean' ) if not isinstance(self.llm_inta_has_fpaa_weight , a ): raise ValueError('llm_int8_has_fp16_weight must be a boolean' ) if self.bnb_abit_compute_dtype is not None and not isinstance(self.bnb_abit_compute_dtype , torch.dtype ): raise ValueError('bnb_4bit_compute_dtype must be torch.dtype' ) if not isinstance(self.bnb_abit_quant_type , a ): raise ValueError('bnb_4bit_quant_type must be a string' ) if not isinstance(self.bnb_abit_use_double_quant , a ): raise ValueError('bnb_4bit_use_double_quant must be a boolean' ) if self.load_in_abit and not version.parse(importlib.metadata.version('bitsandbytes' ) ) >= version.parse( '0.39.0' ): raise ValueError( '4 bit quantization requires bitsandbytes>=0.39.0 - please upgrade your bitsandbytes version' ) def _UpperCamelCase ( self ) -> List[str]: return self.load_in_abit or self.load_in_abit def _UpperCamelCase ( self ) -> Optional[int]: if self.load_in_abit: return "llm_int8" elif self.load_in_abit and self.bnb_abit_quant_type == "fp4": return "fp4" elif self.load_in_abit and self.bnb_abit_quant_type == "nf4": return "nf4" else: return None @classmethod def _UpperCamelCase ( cls , a , a , **a ) -> str: snake_case_ = cls(**a ) snake_case_ = [] for key, value in kwargs.items(): if hasattr(a , a ): setattr(a , a , a ) to_remove.append(a ) for key in to_remove: kwargs.pop(a , a ) if return_unused_kwargs: return config, kwargs else: return config def _UpperCamelCase ( self , a ) -> Dict: with open(a , 'w' , encoding='utf-8' ) as writer: snake_case_ = self.to_dict() snake_case_ = json.dumps(a , indent=2 , sort_keys=a ) + '\n' writer.write(a ) def _UpperCamelCase ( self ) -> Dict[str, Any]: snake_case_ = copy.deepcopy(self.__dict__ ) snake_case_ = str(output['bnb_4bit_compute_dtype'] ).split('.' )[1] return output def __repr__( self ) -> str: return F'''{self.__class__.__name__} {self.to_json_string()}''' def _UpperCamelCase ( self , a = True ) -> str: if use_diff is True: snake_case_ = self.to_diff_dict() else: snake_case_ = self.to_dict() return json.dumps(a , indent=2 , sort_keys=a ) + "\n" def _UpperCamelCase ( self ) -> Dict[str, Any]: snake_case_ = self.to_dict() # get the default config dict snake_case_ = BitsAndBytesConfig().to_dict() snake_case_ = {} # only serialize values that differ from the default config for key, value in config_dict.items(): if value != default_config_dict[key]: snake_case_ = value return serializable_config_dict
178
import collections import importlib.util import os import re from pathlib import Path lowercase = "src/transformers" # Matches is_xxx_available() lowercase = re.compile(r"is\_([a-z_]*)_available()") # Catches a one-line _import_struct = {xxx} lowercase = re.compile(r"^_import_structure\s+=\s+\{([^\}]+)\}") # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowercase = re.compile(r"\s+\"\S*\":\s+\[([^\]]*)\]") # Catches a line if not is_foo_available lowercase = re.compile(r"^\s*if\s+not\s+is\_[a-z_]*\_available\(\)") # Catches a line _import_struct["bla"].append("foo") lowercase = re.compile(r"^\s*_import_structure\[\"\S*\"\]\.append\(\"(\S*)\"\)") # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowercase = re.compile(r"^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]") # Catches a line with an object between quotes and a comma: "MyModel", lowercase = re.compile("^\s+\"([^\"]+)\",") # Catches a line with objects between brackets only: ["foo", "bar"], lowercase = re.compile("^\s+\[([^\]]+)\]") # Catches a line with from foo import bar, bla, boo lowercase = re.compile(r"\s+from\s+\S*\s+import\s+([^\(\s].*)\n") # Catches a line with try: lowercase = re.compile(r"^\s*try:") # Catches a line with else: lowercase = re.compile(r"^\s*else:") def __UpperCAmelCase ( a_): if _re_test_backend.search(a_) is None: return None snake_case_ = [b[0] for b in _re_backend.findall(a_)] backends.sort() return "_and_".join(a_) def __UpperCAmelCase ( a_): with open(a_ , 'r' , encoding='utf-8' , newline='\n') as f: snake_case_ = f.readlines() snake_case_ = 0 while line_index < len(a_) and not lines[line_index].startswith('_import_structure = {'): line_index += 1 # If this is a traditional init, just return. if line_index >= len(a_): return None # First grab the objects without a specific backend in _import_structure snake_case_ = [] while not lines[line_index].startswith('if TYPE_CHECKING') and find_backend(lines[line_index]) is None: snake_case_ = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(a_): snake_case_ = _re_one_line_import_struct.search(a_).groups()[0] snake_case_ = re.findall('\[([^\]]+)\]' , a_) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ')]) line_index += 1 continue snake_case_ = _re_import_struct_key_value.search(a_) if single_line_import_search is not None: snake_case_ = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ') if len(a_) > 0] objects.extend(a_) elif line.startswith(' ' * 8 + '"'): objects.append(line[9:-3]) line_index += 1 snake_case_ = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING'): # If the line is an if not is_backend_available, we grab all objects associated. snake_case_ = find_backend(lines[line_index]) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1]) is None: snake_case_ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index]) is None: line_index += 1 line_index += 1 snake_case_ = [] # Until we unindent, add backend objects to the list while len(lines[line_index]) <= 1 or lines[line_index].startswith(' ' * 4): snake_case_ = lines[line_index] if _re_import_struct_add_one.search(a_) is not None: objects.append(_re_import_struct_add_one.search(a_).groups()[0]) elif _re_import_struct_add_many.search(a_) is not None: snake_case_ = _re_import_struct_add_many.search(a_).groups()[0].split(', ') snake_case_ = [obj[1:-1] for obj in imports if len(a_) > 0] objects.extend(a_) elif _re_between_brackets.search(a_) is not None: snake_case_ = _re_between_brackets.search(a_).groups()[0].split(', ') snake_case_ = [obj[1:-1] for obj in imports if len(a_) > 0] objects.extend(a_) elif _re_quote_object.search(a_) is not None: objects.append(_re_quote_object.search(a_).groups()[0]) elif line.startswith(' ' * 8 + '"'): objects.append(line[9:-3]) elif line.startswith(' ' * 12 + '"'): objects.append(line[13:-3]) line_index += 1 snake_case_ = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend snake_case_ = [] while ( line_index < len(a_) and find_backend(lines[line_index]) is None and not lines[line_index].startswith('else') ): snake_case_ = lines[line_index] snake_case_ = _re_import.search(a_) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ')) elif line.startswith(' ' * 8): objects.append(line[8:-2]) line_index += 1 snake_case_ = {'none': objects} # Let's continue with backend-specific objects while line_index < len(a_): # If the line is an if is_backend_available, we grab all objects associated. snake_case_ = find_backend(lines[line_index]) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1]) is None: snake_case_ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index]) is None: line_index += 1 line_index += 1 snake_case_ = [] # Until we unindent, add backend objects to the list while len(lines[line_index]) <= 1 or lines[line_index].startswith(' ' * 8): snake_case_ = lines[line_index] snake_case_ = _re_import.search(a_) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ')) elif line.startswith(' ' * 12): objects.append(line[12:-2]) line_index += 1 snake_case_ = objects else: line_index += 1 return import_dict_objects, type_hint_objects def __UpperCAmelCase ( a_ , a_): def find_duplicates(a_): return [k for k, v in collections.Counter(a_).items() if v > 1] if list(import_dict_objects.keys()) != list(type_hint_objects.keys()): return ["Both sides of the init do not have the same backends!"] snake_case_ = [] for key in import_dict_objects.keys(): snake_case_ = find_duplicates(import_dict_objects[key]) if duplicate_imports: errors.append(f'''Duplicate _import_structure definitions for: {duplicate_imports}''') snake_case_ = find_duplicates(type_hint_objects[key]) if duplicate_type_hints: errors.append(f'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''') if sorted(set(import_dict_objects[key])) != sorted(set(type_hint_objects[key])): snake_case_ = 'base imports' if key == 'none' else f'''{key} backend''' errors.append(f'''Differences for {name}:''') for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f''' {a} in TYPE_HINT but not in _import_structure.''') for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f''' {a} in _import_structure but not in TYPE_HINT.''') return errors def __UpperCAmelCase ( ): snake_case_ = [] for root, _, files in os.walk(a_): if "__init__.py" in files: snake_case_ = os.path.join(a_ , '__init__.py') snake_case_ = parse_init(a_) if objects is not None: snake_case_ = analyze_results(*a_) if len(a_) > 0: snake_case_ = f'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append('\n'.join(a_)) if len(a_) > 0: raise ValueError('\n\n'.join(a_)) def __UpperCAmelCase ( ): snake_case_ = [] for path, directories, files in os.walk(a_): for folder in directories: # Ignore private modules if folder.startswith('_'): directories.remove(a_) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(a_) / folder).glob('*.py'))) == 0: continue snake_case_ = str((Path(a_) / folder).relative_to(a_)) snake_case_ = short_path.replace(os.path.sep , '.') submodules.append(a_) for fname in files: if fname == "__init__.py": continue snake_case_ = str((Path(a_) / fname).relative_to(a_)) snake_case_ = short_path.replace('.py' , '').replace(os.path.sep , '.') if len(submodule.split('.')) == 1: submodules.append(a_) return submodules lowercase = [ "convert_pytorch_checkpoint_to_tf2", "modeling_flax_pytorch_utils", ] def __UpperCAmelCase ( ): # This is to make sure the transformers module imported is the one in the repo. snake_case_ = importlib.util.spec_from_file_location( 'transformers' , os.path.join(a_ , '__init__.py') , submodule_search_locations=[PATH_TO_TRANSFORMERS] , ) snake_case_ = spec.loader.load_module() snake_case_ = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(a_) > 0: snake_case_ = '\n'.join(f'''- {module}''' for module in module_not_registered) raise ValueError( 'The following submodules are not properly registered in the main init of Transformers:\n' f'''{list_of_modules}\n''' 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.') if __name__ == "__main__": check_all_inits() check_submodules()
178
1
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer a_ : str = logging.get_logger(__name__) a_ : List[Any] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } a_ : Optional[Any] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } a_ : List[Any] = { """facebook/blenderbot_small-90M""": 5_12, } class __UpperCamelCase ( snake_case_ ): lowercase : int =VOCAB_FILES_NAMES lowercase : Any =PRETRAINED_VOCAB_FILES_MAP lowercase : Union[str, Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase : Any =BlenderbotSmallTokenizer def __init__( self, lowerCAmelCase=None, lowerCAmelCase=None, lowerCAmelCase="<|endoftext|>", lowerCAmelCase="<|endoftext|>", lowerCAmelCase="<|endoftext|>", lowerCAmelCase=False, lowerCAmelCase=True, **lowerCAmelCase, ): """simple docstring""" super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase, merges=lowerCAmelCase, add_prefix_space=lowerCAmelCase, trim_offsets=lowerCAmelCase, ), bos_token=lowerCAmelCase, eos_token=lowerCAmelCase, unk_token=lowerCAmelCase, **lowerCAmelCase, ) lowerCamelCase_ =add_prefix_space def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=None ): """simple docstring""" lowerCamelCase_ =[self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase = None ): """simple docstring""" lowerCamelCase_ =[self.sep_token_id] lowerCamelCase_ =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
352
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def a_ ( ) -> Tuple: """simple docstring""" lowerCamelCase_ ={ '''repo_name''': ['''test_repo1''', '''test_repo2''', '''test_repo3'''], '''path''': ['''test_1.py''', '''test_2.py''', '''unit_test.py'''], '''content''': ['''a ''' * 20, '''a ''' * 30, '''b ''' * 7], } lowerCamelCase_ =Dataset.from_dict(__snake_case ) return dataset class __UpperCamelCase ( lowerCamelCase__ ): def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =get_dataset() lowerCamelCase_ =make_duplicate_clusters(lowerCAmelCase, 0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ), 2 ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =get_dataset() lowerCamelCase_, lowerCamelCase_ =deduplicate_dataset(lowerCAmelCase ) self.assertEqual(len(lowerCAmelCase ), 2 ) print(lowerCAmelCase ) self.assertEqual(duplicate_clusters[0][0]['''copies'''], 2 ) self.assertEqual(duplicate_clusters[0][0]['''is_extreme'''], lowerCAmelCase )
6
0
"""simple docstring""" import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _A : List[str] = logging.get_logger(__name__) _A : List[str] = {"""vocab_file""": """spiece.model"""} _A : List[Any] = { """vocab_file""": { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model""", """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model""" ), } } _A : str = { """google/bigbird-roberta-base""": 40_96, """google/bigbird-roberta-large""": 40_96, """google/bigbird-base-trivia-itc""": 40_96, } class a__ ( a_ ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["""input_ids""", """attention_mask"""] __lowerCAmelCase = [] def __init__( self , _a , _a="<unk>" , _a="<s>" , _a="</s>" , _a="<pad>" , _a="[SEP]" , _a="[MASK]" , _a="[CLS]" , _a = None , **_a , ): lowercase : Dict = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else bos_token lowercase : Any = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else eos_token lowercase : Any = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else unk_token lowercase : List[Any] = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else pad_token lowercase : Any = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else cls_token lowercase : Any = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else sep_token # Mask token behave like a normal word, i.e. include the space before it lowercase : List[Any] = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else mask_token lowercase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , pad_token=_a , sep_token=_a , mask_token=_a , cls_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) lowercase : Optional[Any] = vocab_file lowercase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_a ) @property def __magic_name__ ( self ): return self.sp_model.get_piece_size() def __magic_name__ ( self ): lowercase : str = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase : Union[str, Any] = self.__dict__.copy() lowercase : Dict = None return state def __setstate__( self , _a ): lowercase : str = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase : List[str] = {} lowercase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __magic_name__ ( self , _a ): return self.sp_model.encode(_a , out_type=_a ) def __magic_name__ ( self , _a ): return self.sp_model.piece_to_id(_a ) def __magic_name__ ( self , _a ): lowercase : Union[str, Any] = self.sp_model.IdToPiece(_a ) return token def __magic_name__ ( self , _a ): lowercase : List[Any] = [] lowercase : List[Any] = "" lowercase : List[Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_a ) + token lowercase : int = True lowercase : Union[str, Any] = [] else: current_sub_tokens.append(_a ) lowercase : int = False out_string += self.sp_model.decode(_a ) return out_string.strip() def __magic_name__ ( self , _a , _a = False , _a = None , _a = True , **_a , ): lowercase : int = kwargs.pop("use_source_tokenizer" , _a ) lowercase : Union[str, Any] = self.convert_ids_to_tokens(_a , skip_special_tokens=_a ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowercase : Any = [] lowercase : Dict = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(_a ) ) lowercase : int = [] sub_texts.append(_a ) else: current_sub_text.append(_a ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(_a ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: lowercase : Dict = re.sub(R" (\[(MASK|SEP)\])" , R"\1" , " ".join(_a ) ) else: lowercase : Union[str, Any] = "".join(_a ) lowercase : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowercase : Union[str, Any] = self.clean_up_tokenization(_a ) return clean_text else: return text def __magic_name__ ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase : List[Any] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , "wb" ) as fi: lowercase : int = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,) def __magic_name__ ( self , _a , _a = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase : List[str] = [self.cls_token_id] lowercase : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def __magic_name__ ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] def __magic_name__ ( self , _a , _a = None ): lowercase : Tuple = [self.sep_token_id] lowercase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
202
"""simple docstring""" from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker _A : int = """CompVis/stable-diffusion-v1-1""" _A : Any = """CompVis/stable-diffusion-v1-2""" _A : Optional[int] = """CompVis/stable-diffusion-v1-3""" _A : Union[str, Any] = """CompVis/stable-diffusion-v1-4""" class a__ ( a_ ): def __init__( self , _a , _a , _a , _a , _a , _a , _a , _a = True , ): super()._init_() lowercase : Optional[Any] = StableDiffusionPipeline.from_pretrained(_a ) lowercase : str = StableDiffusionPipeline.from_pretrained(_a ) lowercase : Dict = StableDiffusionPipeline.from_pretrained(_a ) lowercase : Union[str, Any] = StableDiffusionPipeline( vae=_a , text_encoder=_a , tokenizer=_a , unet=_a , scheduler=_a , safety_checker=_a , feature_extractor=_a , requires_safety_checker=_a , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def __magic_name__ ( self ): return {k: getattr(self , _a ) for k in self.config.keys() if not k.startswith("_" )} def __magic_name__ ( self , _a = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowercase : str = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_a ) def __magic_name__ ( self ): self.enable_attention_slicing(_a ) @torch.no_grad() def __magic_name__ ( self , _a , _a = 512 , _a = 512 , _a = 50 , _a = 7.5 , _a = None , _a = 1 , _a = 0.0 , _a = None , _a = None , _a = "pil" , _a = True , _a = None , _a = 1 , **_a , ): return self.pipea( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) @torch.no_grad() def __magic_name__ ( self , _a , _a = 512 , _a = 512 , _a = 50 , _a = 7.5 , _a = None , _a = 1 , _a = 0.0 , _a = None , _a = None , _a = "pil" , _a = True , _a = None , _a = 1 , **_a , ): return self.pipea( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) @torch.no_grad() def __magic_name__ ( self , _a , _a = 512 , _a = 512 , _a = 50 , _a = 7.5 , _a = None , _a = 1 , _a = 0.0 , _a = None , _a = None , _a = "pil" , _a = True , _a = None , _a = 1 , **_a , ): return self.pipea( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) @torch.no_grad() def __magic_name__ ( self , _a , _a = 512 , _a = 512 , _a = 50 , _a = 7.5 , _a = None , _a = 1 , _a = 0.0 , _a = None , _a = None , _a = "pil" , _a = True , _a = None , _a = 1 , **_a , ): return self.pipea( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) @torch.no_grad() def __magic_name__ ( self , _a , _a = 512 , _a = 512 , _a = 50 , _a = 7.5 , _a = None , _a = 1 , _a = 0.0 , _a = None , _a = None , _a = "pil" , _a = True , _a = None , _a = 1 , **_a , ): lowercase : List[Any] = "cuda" if torch.cuda.is_available() else "cpu" self.to(_a ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` must be divisible by 8 but are {height} and {width}.""" ) # Get first result from Stable Diffusion Checkpoint v1.1 lowercase : List[Any] = self.textaimg_sda_a( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) # Get first result from Stable Diffusion Checkpoint v1.2 lowercase : Any = self.textaimg_sda_a( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) # Get first result from Stable Diffusion Checkpoint v1.3 lowercase : str = self.textaimg_sda_a( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) # Get first result from Stable Diffusion Checkpoint v1.4 lowercase : Optional[int] = self.textaimg_sda_a( prompt=_a , height=_a , width=_a , num_inference_steps=_a , guidance_scale=_a , negative_prompt=_a , num_images_per_prompt=_a , eta=_a , generator=_a , latents=_a , output_type=_a , return_dict=_a , callback=_a , callback_steps=_a , **_a , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
202
1
def lowerCAmelCase_ ( snake_case_ ): if not isinstance(snake_case_,snake_case_ ): _A : List[str] = f'''Input value of [number={number}] must be an integer''' raise TypeError(snake_case_ ) if number < 1: _A : Dict = f'''Input value of [number={number}] must be > 0''' raise ValueError(snake_case_ ) _A : List[str] = 1 for i in range(1,snake_case_ ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
343
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class lowercase ( unittest.TestCase ): def a__ ( self ) -> List[str]: debug_launcher(test_script.main ) def a__ ( self ) -> Any: debug_launcher(test_ops.main )
343
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): @slow def a__ ( self ) -> Any: _A : Tuple = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) _A : List[Any] = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 2_5543, 110, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _A : List[str] = model(_a )["""last_hidden_state"""] _A : Union[str, Any] = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape , _a ) # compare the actual values for a slice. _A : List[Any] = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
26
'''simple docstring''' def _A ( snake_case , snake_case ) -> float: return price * (1 + tax_rate) if __name__ == "__main__": print(F'''{price_plus_tax(100, 0.2_5) = }''') print(F'''{price_plus_tax(1_2_5.5_0, 0.0_5) = }''')
250
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a = logging.get_logger(__name__) _a = { """microsoft/swin-tiny-patch4-window7-224""": ( """https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json""" ), # See all Swin models at https://huggingface.co/models?filter=swin } class _UpperCAmelCase( lowerCamelCase , lowerCamelCase ): lowercase__ = 'swin' lowercase__ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , __a=2_24 , __a=4 , __a=3 , __a=96 , __a=[2, 2, 6, 2] , __a=[3, 6, 12, 24] , __a=7 , __a=4.0 , __a=True , __a=0.0 , __a=0.0 , __a=0.1 , __a="gelu" , __a=False , __a=0.02 , __a=1e-5 , __a=32 , __a=None , __a=None , **__a , ) -> List[str]: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = embed_dim _UpperCamelCase = depths _UpperCamelCase = len(__a) _UpperCamelCase = num_heads _UpperCamelCase = window_size _UpperCamelCase = mlp_ratio _UpperCamelCase = qkv_bias _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = drop_path_rate _UpperCamelCase = hidden_act _UpperCamelCase = use_absolute_embeddings _UpperCamelCase = layer_norm_eps _UpperCamelCase = initializer_range _UpperCamelCase = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _UpperCamelCase = int(embed_dim * 2 ** (len(__a) - 1)) _UpperCamelCase = ['''stem'''] + [F'''stage{idx}''' for idx in range(1 , len(__a) + 1)] _UpperCamelCase , _UpperCamelCase = get_aligned_output_features_output_indices( out_features=__a , out_indices=__a , stage_names=self.stage_names) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = version.parse('1.11' ) @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ]) @property def UpperCAmelCase ( self) -> float: '''simple docstring''' return 1e-4
354
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { """facebook/detr-resnet-50""": """https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json""", # See all DETR models at https://huggingface.co/models?filter=detr } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'detr' lowercase__ = ['past_key_values'] lowercase__ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , __a=True , __a=None , __a=3 , __a=1_00 , __a=6 , __a=20_48 , __a=8 , __a=6 , __a=20_48 , __a=8 , __a=0.0 , __a=0.0 , __a=True , __a="relu" , __a=2_56 , __a=0.1 , __a=0.0 , __a=0.0 , __a=0.02 , __a=1.0 , __a=False , __a="sine" , __a="resnet50" , __a=True , __a=False , __a=1 , __a=5 , __a=2 , __a=1 , __a=1 , __a=5 , __a=2 , __a=0.1 , **__a , ) -> Tuple: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''') if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''') _UpperCamelCase = CONFIG_MAPPING['''resnet'''](out_features=['''stage4''']) elif isinstance(__a , __a): _UpperCamelCase = backbone_config.get('''model_type''') _UpperCamelCase = CONFIG_MAPPING[backbone_model_type] _UpperCamelCase = config_class.from_dict(__a) # set timm attributes to None _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None, None, None _UpperCamelCase = use_timm_backbone _UpperCamelCase = backbone_config _UpperCamelCase = num_channels _UpperCamelCase = num_queries _UpperCamelCase = d_model _UpperCamelCase = encoder_ffn_dim _UpperCamelCase = encoder_layers _UpperCamelCase = encoder_attention_heads _UpperCamelCase = decoder_ffn_dim _UpperCamelCase = decoder_layers _UpperCamelCase = decoder_attention_heads _UpperCamelCase = dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = activation_function _UpperCamelCase = init_std _UpperCamelCase = init_xavier_std _UpperCamelCase = encoder_layerdrop _UpperCamelCase = decoder_layerdrop _UpperCamelCase = encoder_layers _UpperCamelCase = auxiliary_loss _UpperCamelCase = position_embedding_type _UpperCamelCase = backbone _UpperCamelCase = use_pretrained_backbone _UpperCamelCase = dilation # Hungarian matcher _UpperCamelCase = class_cost _UpperCamelCase = bbox_cost _UpperCamelCase = giou_cost # Loss coefficients _UpperCamelCase = mask_loss_coefficient _UpperCamelCase = dice_loss_coefficient _UpperCamelCase = bbox_loss_coefficient _UpperCamelCase = giou_loss_coefficient _UpperCamelCase = eos_coefficient super().__init__(is_encoder_decoder=__a , **__a) @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self.encoder_attention_heads @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self.d_model @classmethod def UpperCAmelCase ( cls , __a , **__a) -> Any: '''simple docstring''' return cls(backbone_config=__a , **__a) def UpperCAmelCase ( self) -> Dict[str, any]: '''simple docstring''' _UpperCamelCase = copy.deepcopy(self.__dict__) if output["backbone_config"] is not None: _UpperCamelCase = self.backbone_config.to_dict() _UpperCamelCase = self.__class__.model_type return output class _UpperCAmelCase( lowerCamelCase ): lowercase__ = version.parse('1.11' ) @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ]) @property def UpperCAmelCase ( self) -> float: '''simple docstring''' return 1e-5 @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 12
100
0
'''simple docstring''' import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( '''The `image_to_image.py` script is outdated. Please use directly `from diffusers import''' ''' StableDiffusionImg2ImgPipeline` instead.''' )
53
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class __snake_case ( lowerCAmelCase ): _a : Union[str, Any]= "microsoft/speecht5_tts" _a : Tuple= ( "This is a tool that reads an English text out loud. It takes an input named `text` which should contain the " "text to read (in English) and returns a waveform object containing the sound." ) _a : Dict= "text_reader" _a : Optional[Any]= SpeechTaProcessor _a : Tuple= SpeechTaForTextToSpeech _a : Optional[int]= SpeechTaHifiGan _a : Union[str, Any]= ["text"] _a : Optional[int]= ["audio"] def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' if self.post_processor is None: lowercase : Any = """microsoft/speecht5_hifigan""" super().setup() def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case=None ): '''simple docstring''' lowercase : int = self.pre_processor(text=snake_case ,return_tensors="""pt""" ,truncation=snake_case ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError("""Datasets needs to be installed if not passing speaker embeddings.""" ) lowercase : Tuple = load_dataset("""Matthijs/cmu-arctic-xvectors""" ,split="""validation""" ) lowercase : List[str] = torch.tensor(embeddings_dataset[7305]["""xvector"""] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' with torch.no_grad(): return self.model.generate_speech(**snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' with torch.no_grad(): return self.post_processor(snake_case ).cpu().detach()
20
0
import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _snake_case ( unittest.TestCase ): '''simple docstring''' def A__ ( self: Optional[Any] ) -> int: UpperCAmelCase_ : Dict = tempfile.mkdtemp() # fmt: off UpperCAmelCase_ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on UpperCAmelCase_ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) UpperCAmelCase_ : List[str] = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } UpperCAmelCase_ : Union[str, Any] = os.path.join(self.tmpdirname ,lowerCamelCase_ ) with open(self.image_processor_file ,"""w""" ,encoding="""utf-8""" ) as fp: json.dump(lowerCamelCase_ ,lowerCamelCase_ ) def A__ ( self: List[str] ,**lowerCamelCase_: int ) -> Union[str, Any]: return BertTokenizer.from_pretrained(self.tmpdirname ,**lowerCamelCase_ ) def A__ ( self: Any ,**lowerCamelCase_: int ) -> str: return ViTImageProcessor.from_pretrained(self.tmpdirname ,**lowerCamelCase_ ) def A__ ( self: Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def A__ ( self: Dict ) -> Tuple: UpperCAmelCase_ : List[Any] = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] UpperCAmelCase_ : List[str] = [Image.fromarray(np.moveaxis(lowerCamelCase_ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def A__ ( self: Optional[Any] ) -> Tuple: UpperCAmelCase_ : int = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = self.get_image_processor() UpperCAmelCase_ : Union[str, Any] = VisionTextDualEncoderProcessor(tokenizer=lowerCamelCase_ ,image_processor=lowerCamelCase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Dict = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,(BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase_ ) def A__ ( self: Tuple ) -> Optional[int]: UpperCAmelCase_ : Any = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : List[str] = self.get_tokenizer(bos_token="""(BOS)""" ,eos_token="""(EOS)""" ) UpperCAmelCase_ : List[str] = self.get_image_processor(do_normalize=lowerCamelCase_ ,padding_value=1.0 ) UpperCAmelCase_ : Optional[Any] = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname ,bos_token="""(BOS)""" ,eos_token="""(EOS)""" ,do_normalize=lowerCamelCase_ ,padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer ,(BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,lowerCamelCase_ ) def A__ ( self: Tuple ) -> Optional[int]: UpperCAmelCase_ : Tuple = self.get_image_processor() UpperCAmelCase_ : Dict = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=lowerCamelCase_ ,image_processor=lowerCamelCase_ ) UpperCAmelCase_ : int = self.prepare_image_inputs() UpperCAmelCase_ : List[str] = image_processor(lowerCamelCase_ ,return_tensors="""np""" ) UpperCAmelCase_ : Union[str, Any] = processor(images=lowerCamelCase_ ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def A__ ( self: Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase_ : int = self.get_image_processor() UpperCAmelCase_ : List[str] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=lowerCamelCase_ ,image_processor=lowerCamelCase_ ) UpperCAmelCase_ : Tuple = """lower newer""" UpperCAmelCase_ : Tuple = processor(text=lowerCamelCase_ ) UpperCAmelCase_ : Any = tokenizer(lowerCamelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def A__ ( self: List[str] ) -> str: UpperCAmelCase_ : List[Any] = self.get_image_processor() UpperCAmelCase_ : Optional[Any] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=lowerCamelCase_ ,image_processor=lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = """lower newer""" UpperCAmelCase_ : List[str] = self.prepare_image_inputs() UpperCAmelCase_ : List[Any] = processor(text=lowerCamelCase_ ,images=lowerCamelCase_ ) self.assertListEqual(list(inputs.keys() ) ,["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(lowerCamelCase_ ): processor() def A__ ( self: Optional[int] ) -> Union[str, Any]: UpperCAmelCase_ : Optional[int] = self.get_image_processor() UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : List[str] = VisionTextDualEncoderProcessor(tokenizer=lowerCamelCase_ ,image_processor=lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase_ : Optional[int] = processor.batch_decode(lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = tokenizer.batch_decode(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ ,lowerCamelCase_ ) def A__ ( self: Any ) -> Dict: UpperCAmelCase_ : int = self.get_image_processor() UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : str = VisionTextDualEncoderProcessor(tokenizer=lowerCamelCase_ ,image_processor=lowerCamelCase_ ) UpperCAmelCase_ : int = """lower newer""" UpperCAmelCase_ : str = self.prepare_image_inputs() UpperCAmelCase_ : str = processor(text=lowerCamelCase_ ,images=lowerCamelCase_ ) self.assertListEqual(list(inputs.keys() ) ,processor.model_input_names )
368
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _snake_case ( __snake_case ): '''simple docstring''' A__ : Optional[int] = "roformer" def __init__( self: Optional[int] ,lowerCamelCase_: Tuple=50000 ,lowerCamelCase_: Optional[int]=None ,lowerCamelCase_: List[Any]=768 ,lowerCamelCase_: List[Any]=12 ,lowerCamelCase_: Optional[int]=12 ,lowerCamelCase_: Optional[Any]=3072 ,lowerCamelCase_: int="gelu" ,lowerCamelCase_: str=0.1 ,lowerCamelCase_: Union[str, Any]=0.1 ,lowerCamelCase_: Any=1536 ,lowerCamelCase_: str=2 ,lowerCamelCase_: Optional[int]=0.0_2 ,lowerCamelCase_: int=1e-12 ,lowerCamelCase_: Optional[int]=0 ,lowerCamelCase_: Any=False ,lowerCamelCase_: Union[str, Any]=True ,**lowerCamelCase_: List[str] ,) -> Tuple: super().__init__(pad_token_id=lowerCamelCase_ ,**lowerCamelCase_ ) UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Optional[int] = hidden_size if embedding_size is None else embedding_size UpperCAmelCase_ : Optional[Any] = hidden_size UpperCAmelCase_ : Optional[int] = num_hidden_layers UpperCAmelCase_ : Any = num_attention_heads UpperCAmelCase_ : Optional[Any] = hidden_act UpperCAmelCase_ : Any = intermediate_size UpperCAmelCase_ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase_ : Optional[int] = attention_probs_dropout_prob UpperCAmelCase_ : Tuple = max_position_embeddings UpperCAmelCase_ : Optional[Any] = type_vocab_size UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : Optional[Any] = rotary_value UpperCAmelCase_ : str = use_cache class _snake_case ( __snake_case ): '''simple docstring''' @property def A__ ( self: Tuple ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCAmelCase_ : Tuple = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase_ : Optional[Any] = {0: """batch""", 1: """sequence"""} UpperCAmelCase_ : Any = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
59
0
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _lowerCAmelCase : int = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class __magic_name__ ( unittest.TestCase ): def __magic_name__ ( self , __snake_case , __snake_case = None , __snake_case = None , __snake_case = None , __snake_case = True , ) -> Any: '''simple docstring''' __a =[file for file in os.listdir(__snake_case ) if os.path.isfile(os.path.join(__snake_case , __snake_case ) )] if identifier is not None: __a =[file for file in files if identifier in file] if n_identifier is not None: if isinstance(__snake_case , __snake_case ): for n_ in n_identifier: __a =[file for file in files if n_ not in file] else: __a =[file for file in files if n_identifier not in file] __a =ignore_files or [] ignore_files.append('__init__.py' ) __a =[file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' , __snake_case ) if only_modules: __a =file.split('.' )[0] try: __a =getattr(__snake_case , __snake_case ) __a =doctest.DocTestSuite(__snake_case ) __a =unittest.TextTestRunner().run(__snake_case ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(f'{module_identifier} is not a module.' ) else: __a =doctest.testfile(str('..' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def __magic_name__ ( self ) -> Tuple: '''simple docstring''' __a =Path('src/transformers' ) __a ='modeling' __a =[ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(__snake_case , identifier=__snake_case , ignore_files=__snake_case ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =Path('src/transformers' ) __a ='tokenization' self.analyze_directory(__snake_case , identifier=__snake_case ) def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =Path('src/transformers' ) __a ='configuration' self.analyze_directory(__snake_case , identifier=__snake_case ) def __magic_name__ ( self ) -> List[str]: '''simple docstring''' __a =Path('src/transformers' ) __a =['configuration', 'modeling', 'tokenization'] self.analyze_directory(__snake_case , n_identifier=__snake_case ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =Path('docs/source' ) __a =['favicon.ico'] self.analyze_directory(__snake_case , ignore_files=__snake_case , only_modules=__snake_case )
218
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
0
"""simple docstring""" from math import factorial, radians def a__ ( lowerCAmelCase , lowerCAmelCase = 18 , lowerCAmelCase = 10 ) -> float: UpperCAmelCase__ : List[Any] = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians UpperCAmelCase__ : Any = radians(lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = angle_in_radians UpperCAmelCase__ : Tuple = 3 UpperCAmelCase__ : str = -1 for _ in range(lowerCAmelCase ): result += (b * (angle_in_radians**a)) / factorial(lowerCAmelCase ) UpperCAmelCase__ : int = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(lowerCAmelCase , lowerCAmelCase ) if __name__ == "__main__": __import__("""doctest""").testmod()
166
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=4 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[2, 2, 3, 2] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=["stage2", "stage3", "stage4"] , _lowerCamelCase=3 , _lowerCamelCase=None , ): """simple docstring""" UpperCAmelCase__ : int = parent UpperCAmelCase__ : str = batch_size UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : Optional[Any] = num_channels UpperCAmelCase__ : Optional[int] = num_stages UpperCAmelCase__ : Optional[Any] = hidden_sizes UpperCAmelCase__ : Any = depths UpperCAmelCase__ : str = is_training UpperCAmelCase__ : Tuple = use_labels UpperCAmelCase__ : Optional[Any] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : Tuple = type_sequence_label_size UpperCAmelCase__ : Dict = initializer_range UpperCAmelCase__ : Tuple = out_features UpperCAmelCase__ : Dict = num_labels UpperCAmelCase__ : Tuple = scope UpperCAmelCase__ : Optional[int] = num_stages def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : List[Any] = None if self.use_labels: UpperCAmelCase__ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Dict = self.get_config() return config, pixel_values, labels def _a (self ): """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def _a (self ): """simple docstring""" return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=_lowerCamelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=_lowerCamelCase , loss_ignore_index=255 , num_labels=self.num_labels , ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Any = UperNetForSemanticSegmentation(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Tuple = model(_lowerCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Union[str, Any] = config_and_inputs UpperCAmelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = (UperNetForSemanticSegmentation,) if is_torch_available() else () SCREAMING_SNAKE_CASE = {'image-segmentation': UperNetForSemanticSegmentation} if is_torch_available() else {} SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = UperNetModelTester(self ) UpperCAmelCase__ : Dict = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def _a (self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a (self ): """simple docstring""" return def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class(_lowerCamelCase ) UpperCAmelCase__ : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : List[Any] = [*signature.parameters.keys()] UpperCAmelCase__ : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_lowerCamelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""UperNet does not have a base model""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""UperNet does not have a base model""" ) def _a (self ): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def _a (self ): """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _a (self ): """simple docstring""" pass def _a (self ): """simple docstring""" def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): UpperCAmelCase__ : List[Any] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): UpperCAmelCase__ : int = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) UpperCAmelCase__ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase__ : Tuple = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase__ : Any = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : List[Any] = _config_zero_init(_lowerCamelCase ) UpperCAmelCase__ : List[str] = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class(config=_lowerCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def _a (self ): """simple docstring""" pass @slow def _a (self ): """simple docstring""" for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : str = UperNetForSemanticSegmentation.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def a__ ( ) -> List[Any]: UpperCAmelCase__ : List[str] = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) UpperCAmelCase__ : str = Image.open(lowerCAmelCase ).convert("""RGB""" ) return image @require_torch @require_vision @slow class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) UpperCAmelCase__ : Any = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(_lowerCamelCase ) UpperCAmelCase__ : Optional[Any] = prepare_img() UpperCAmelCase__ : int = processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) with torch.no_grad(): UpperCAmelCase__ : Any = model(**_lowerCamelCase ) UpperCAmelCase__ : List[Any] = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) UpperCAmelCase__ : Optional[Any] = torch.tensor( [[-7.5_958, -7.5_958, -7.4_302], [-7.5_958, -7.5_958, -7.4_302], [-7.4_797, -7.4_797, -7.3_068]] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , _lowerCamelCase , atol=1e-4 ) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) UpperCAmelCase__ : List[str] = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(_lowerCamelCase ) UpperCAmelCase__ : Tuple = prepare_img() UpperCAmelCase__ : int = processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) with torch.no_grad(): UpperCAmelCase__ : int = model(**_lowerCamelCase ) UpperCAmelCase__ : str = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) UpperCAmelCase__ : int = torch.tensor( [[-8.8_110, -8.8_110, -8.6_521], [-8.8_110, -8.8_110, -8.6_521], [-8.7_746, -8.7_746, -8.6_130]] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , _lowerCamelCase , atol=1e-4 ) )
166
1
"""simple docstring""" from collections.abc import Sequence def _lowerCAmelCase ( lowercase_ , lowercase_ = False ): if not arr: return 0 UpperCAmelCase = 0 if allow_empty_subarrays else float('-inf' ) UpperCAmelCase = 0.0 for num in arr: UpperCAmelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) UpperCAmelCase = max(lowercase_ , lowercase_ ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() snake_case_ = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(f'''{max_subarray_sum(nums) = }''')
78
import pytest __UpperCAmelCase : Optional[Any] = "__dummy_dataset1__" __UpperCAmelCase : List[str] = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def A__ ( ) -> Optional[int]: return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def A__ ( ) -> Tuple: return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) -> Tuple: __snake_case: List[Any] = dataset_loading_script_name __snake_case: Any = tmp_path / """datasets""" / script_name script_dir.mkdir(parents=SCREAMING_SNAKE_CASE__) __snake_case: int = script_dir / F'''{script_name}.py''' with open(SCREAMING_SNAKE_CASE__ , """w""") as f: f.write(SCREAMING_SNAKE_CASE__) return str(SCREAMING_SNAKE_CASE__)
111
0
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase = "" , UpperCAmelCase = False ) -> None: _snake_case = {} # A node will be a leaf if the tree contains its word _snake_case = is_leaf _snake_case = prefix def lowercase (self , UpperCAmelCase ) -> tuple[str, str, str]: _snake_case = 0 for q, w in zip(self.prefix , lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def lowercase (self , UpperCAmelCase ) -> None: for word in words: self.insert(lowerCAmelCase__ ) def lowercase (self , UpperCAmelCase ) -> None: if self.prefix == word: _snake_case = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _snake_case = RadixNode(prefix=lowerCAmelCase__ , is_leaf=lowerCAmelCase__ ) else: _snake_case = self.nodes[word[0]] _snake_case = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _snake_case = remaining_prefix _snake_case = self.nodes[matching_string[0]] _snake_case = RadixNode(lowerCAmelCase__ , lowerCAmelCase__ ) _snake_case = aux_node if remaining_word == "": _snake_case = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def lowercase (self , UpperCAmelCase ) -> bool: _snake_case = self.nodes.get(word[0] , lowerCAmelCase__ ) if not incoming_node: return False else: _snake_case = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def lowercase (self , UpperCAmelCase ) -> bool: _snake_case = self.nodes.get(word[0] , lowerCAmelCase__ ) if not incoming_node: return False else: _snake_case = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _snake_case = list(self.nodes.values() )[0] _snake_case = merging_node.is_leaf self.prefix += merging_node.prefix _snake_case = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _snake_case = False # If there is 1 edge, we merge it with its child else: _snake_case = list(incoming_node.nodes.values() )[0] _snake_case = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _snake_case = merging_node.nodes return True def lowercase (self , UpperCAmelCase = 0 ) -> None: if self.prefix != "": print("""-""" * height , self.prefix , """ (leaf)""" if self.is_leaf else """""" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = "banana bananas bandana band apple all beast".split() _snake_case = RadixNode() root.insert_many(a_ ) assert all(root.find(a_ ) for word in words ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def __SCREAMING_SNAKE_CASE ( ): assert test_trie() def __SCREAMING_SNAKE_CASE ( ): _snake_case = RadixNode() _snake_case = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(a_ ) print("""Words:""" , a_ ) print("""Tree:""" ) root.print_tree() if __name__ == "__main__": main()
367
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCAmelCase = { 'configuration_resnet': ['RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ResNetConfig', 'ResNetOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'ResNetForImageClassification', 'ResNetModel', 'ResNetPreTrainedModel', 'ResNetBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFResNetForImageClassification', 'TFResNetModel', 'TFResNetPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FlaxResNetForImageClassification', 'FlaxResNetModel', 'FlaxResNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
270
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class A__ ( metaclass=UpperCamelCase ): """simple docstring""" UpperCamelCase_ : List[Any] = ['''flax''', '''transformers'''] def __init__( self : Dict , *lowerCAmelCase__ : str , **lowerCAmelCase__ : Union[str, Any] ) -> int: """simple docstring""" requires_backends(self , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : List[Any] , *lowerCAmelCase__ : int , **lowerCAmelCase__ : Union[str, Any] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : Optional[int] , *lowerCAmelCase__ : Optional[int] , **lowerCAmelCase__ : Optional[int] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) class A__ ( metaclass=UpperCamelCase ): """simple docstring""" UpperCamelCase_ : Dict = ['''flax''', '''transformers'''] def __init__( self : str , *lowerCAmelCase__ : List[Any] , **lowerCAmelCase__ : Tuple ) -> Tuple: """simple docstring""" requires_backends(self , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : Union[str, Any] , *lowerCAmelCase__ : List[Any] , **lowerCAmelCase__ : Dict ) -> List[str]: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : List[str] , *lowerCAmelCase__ : str , **lowerCAmelCase__ : List[Any] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) class A__ ( metaclass=UpperCamelCase ): """simple docstring""" UpperCamelCase_ : Tuple = ['''flax''', '''transformers'''] def __init__( self : Any , *lowerCAmelCase__ : List[str] , **lowerCAmelCase__ : Tuple ) -> str: """simple docstring""" requires_backends(self , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : Any , *lowerCAmelCase__ : Tuple , **lowerCAmelCase__ : int ) -> str: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : Optional[Any] , *lowerCAmelCase__ : str , **lowerCAmelCase__ : List[str] ) -> int: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) class A__ ( metaclass=UpperCamelCase ): """simple docstring""" UpperCamelCase_ : List[Any] = ['''flax''', '''transformers'''] def __init__( self : Optional[int] , *lowerCAmelCase__ : Dict , **lowerCAmelCase__ : Tuple ) -> Tuple: """simple docstring""" requires_backends(self , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : Optional[Any] , *lowerCAmelCase__ : List[Any] , **lowerCAmelCase__ : Optional[int] ) -> int: """simple docstring""" requires_backends(cls , ["flax", "transformers"] ) @classmethod def _lowerCAmelCase ( cls : Tuple , *lowerCAmelCase__ : Optional[int] , **lowerCAmelCase__ : Tuple ) -> int: """simple docstring""" requires_backends(cls , ["flax", "transformers"] )
145
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) # TODO Update this UpperCAmelCase__ = { '''facebook/esm-1b''': '''https://huggingface.co/facebook/esm-1b/resolve/main/config.json''', # See all ESM models at https://huggingface.co/models?filter=esm } class lowerCamelCase__ ( lowerCAmelCase): SCREAMING_SNAKE_CASE__ = '''esm''' def __init__(self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=7_6_8 , UpperCAmelCase=1_2 , UpperCAmelCase=1_2 , UpperCAmelCase=3_0_7_2 , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=1_0_2_6 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-12 , UpperCAmelCase="absolute" , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=False , UpperCAmelCase=False , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase , ) -> Tuple: super().__init__(pad_token_id=UpperCAmelCase , mask_token_id=UpperCAmelCase , **UpperCAmelCase ) _lowercase =vocab_size _lowercase =hidden_size _lowercase =num_hidden_layers _lowercase =num_attention_heads _lowercase =intermediate_size _lowercase =hidden_dropout_prob _lowercase =attention_probs_dropout_prob _lowercase =max_position_embeddings _lowercase =initializer_range _lowercase =layer_norm_eps _lowercase =position_embedding_type _lowercase =use_cache _lowercase =emb_layer_norm_before _lowercase =token_dropout _lowercase =is_folding_model if is_folding_model: if esmfold_config is None: logger.info('''No esmfold_config supplied for folding model, using default values.''' ) _lowercase =EsmFoldConfig() elif isinstance(UpperCAmelCase , UpperCAmelCase ): _lowercase =EsmFoldConfig(**UpperCAmelCase ) _lowercase =esmfold_config if vocab_list is None: logger.warning('''No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!''' ) _lowercase =get_default_vocab_list() else: _lowercase =vocab_list else: _lowercase =None _lowercase =None if self.esmfold_config is not None and getattr(self.esmfold_config , '''use_esm_attn_map''' , UpperCAmelCase ): raise ValueError('''The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!''' ) def __A (self ) -> List[str]: _lowercase =super().to_dict() if isinstance(self.esmfold_config , UpperCAmelCase ): _lowercase =self.esmfold_config.to_dict() return output @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = None def __A (self ) -> Union[str, Any]: if self.trunk is None: _lowercase =TrunkConfig() elif isinstance(self.trunk , UpperCAmelCase ): _lowercase =TrunkConfig(**self.trunk ) def __A (self ) -> Tuple: _lowercase =asdict(self ) _lowercase =self.trunk.to_dict() return output @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE__ = 48 SCREAMING_SNAKE_CASE__ = 1024 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 4 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = None def __A (self ) -> List[str]: if self.structure_module is None: _lowercase =StructureModuleConfig() elif isinstance(self.structure_module , UpperCAmelCase ): _lowercase =StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( '''`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got''' f" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( '''`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got''' f" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) _lowercase =self.sequence_state_dim // self.sequence_head_width _lowercase =self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( '''`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got''' f" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( '''`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got''' f" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(f"`dropout` should not be greater than 0.4, got {self.dropout}." ) def __A (self ) -> Dict: _lowercase =asdict(self ) _lowercase =self.structure_module.to_dict() return output @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE__ = 384 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 16 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 12 SCREAMING_SNAKE_CASE__ = 4 SCREAMING_SNAKE_CASE__ = 8 SCREAMING_SNAKE_CASE__ = 0.1 SCREAMING_SNAKE_CASE__ = 8 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = 7 SCREAMING_SNAKE_CASE__ = 10 SCREAMING_SNAKE_CASE__ = 1E-8 SCREAMING_SNAKE_CASE__ = 1E5 def __A (self ) -> List[Any]: return asdict(self ) def UpperCAmelCase_ ( ) -> Tuple: """simple docstring""" return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
5
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): A_ : Optional[int] = SwinvaConfig() A_ : str = swinva_name.split('''_''' ) A_ : Any = name_split[1] if "to" in name_split[3]: A_ : List[str] = int(name_split[3][-3:] ) else: A_ : List[str] = int(name_split[3] ) if "to" in name_split[2]: A_ : List[Any] = int(name_split[2][-2:] ) else: A_ : str = int(name_split[2][6:] ) if model_size == "tiny": A_ : Tuple = 96 A_ : str = (2, 2, 6, 2) A_ : List[str] = (3, 6, 12, 24) elif model_size == "small": A_ : Optional[Any] = 96 A_ : int = (2, 2, 18, 2) A_ : str = (3, 6, 12, 24) elif model_size == "base": A_ : int = 128 A_ : Optional[int] = (2, 2, 18, 2) A_ : int = (4, 8, 16, 32) else: A_ : Dict = 192 A_ : Any = (2, 2, 18, 2) A_ : str = (6, 12, 24, 48) if "to" in swinva_name: A_ : List[Any] = (12, 12, 12, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): A_ : Union[str, Any] = 21_841 A_ : List[Any] = 'huggingface/label-files' A_ : str = 'imagenet-22k-id2label.json' A_ : Union[str, Any] = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type='''dataset''' ) , '''r''' ) ) A_ : int = {int(snake_case__ ): v for k, v in idalabel.items()} A_ : Dict = idalabel A_ : Union[str, Any] = {v: k for k, v in idalabel.items()} else: A_ : Dict = 1_000 A_ : str = 'huggingface/label-files' A_ : Optional[Any] = 'imagenet-1k-id2label.json' A_ : Any = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type='''dataset''' ) , '''r''' ) ) A_ : str = {int(snake_case__ ): v for k, v in idalabel.items()} A_ : Optional[Any] = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} A_ : Any = img_size A_ : Any = num_classes A_ : Optional[int] = embed_dim A_ : Optional[Any] = depths A_ : Optional[int] = num_heads A_ : Optional[Any] = window_size return config def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): if "patch_embed.proj" in name: A_ : Tuple = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: A_ : List[str] = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: A_ : List[Any] = 'encoder.' + name if "attn.proj" in name: A_ : Union[str, Any] = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: A_ : str = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: A_ : List[Any] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: A_ : Dict = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: A_ : List[str] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: A_ : str = name.replace('''mlp.fc2''' , '''output.dense''' ) if "q_bias" in name: A_ : int = name.replace('''q_bias''' , '''query.bias''' ) if "k_bias" in name: A_ : Tuple = name.replace('''k_bias''' , '''key.bias''' ) if "v_bias" in name: A_ : Optional[Any] = name.replace('''v_bias''' , '''value.bias''' ) if "cpb_mlp" in name: A_ : Optional[Any] = name.replace('''cpb_mlp''' , '''continuous_position_bias_mlp''' ) if name == "norm.weight": A_ : Optional[int] = 'layernorm.weight' if name == "norm.bias": A_ : List[str] = 'layernorm.bias' if "head" in name: A_ : Optional[Any] = name.replace('''head''' , '''classifier''' ) else: A_ : Optional[int] = 'swinv2.' + name return name def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): for key in orig_state_dict.copy().keys(): A_ : Any = orig_state_dict.pop(snake_case__ ) if "mask" in key: continue elif "qkv" in key: A_ : Tuple = key.split('''.''' ) A_ : Tuple = int(key_split[1] ) A_ : Any = int(key_split[3] ) A_ : Any = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: A_ : List[str] = val[:dim, :] A_ : str = val[dim : dim * 2, :] A_ : str = val[-dim:, :] else: A_ : int = val[:dim] A_ : Any = val[ dim : dim * 2 ] A_ : Optional[int] = val[-dim:] else: A_ : Union[str, Any] = val return orig_state_dict def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : Dict = timm.create_model(snake_case__ , pretrained=snake_case__ ) timm_model.eval() A_ : int = get_swinva_config(snake_case__ ) A_ : List[str] = SwinvaForImageClassification(snake_case__ ) model.eval() A_ : Any = convert_state_dict(timm_model.state_dict() , snake_case__ ) model.load_state_dict(snake_case__ ) A_ : List[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' A_ : Dict = AutoImageProcessor.from_pretrained('''microsoft/{}'''.format(swinva_name.replace('''_''' , '''-''' ) ) ) A_ : Optional[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) A_ : Tuple = image_processor(images=snake_case__ , return_tensors='''pt''' ) A_ : Optional[Any] = timm_model(inputs['''pixel_values'''] ) A_ : Union[str, Any] = model(**snake_case__ ).logits assert torch.allclose(snake_case__ , snake_case__ , atol=1e-3 ) print(f'''Saving model {swinva_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(snake_case__ ) model.push_to_hub( repo_path_or_name=Path(snake_case__ , snake_case__ ) , organization='''nandwalritik''' , commit_message='''Add model''' , ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--swinv2_name""", default="""swinv2_tiny_patch4_window8_256""", type=str, help="""Name of the Swinv2 timm model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) UpperCamelCase = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
364
import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class _lowerCamelCase ( unittest.TestCase ): """simple docstring""" def _snake_case ( self )->Any: '''simple docstring''' A_ : Dict = '''hf-internal-testing/tiny-random-t5''' A_ : str = AutoTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ) A_ : Union[str, Any] = AutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE ) A_ : Union[str, Any] = tokenizer('''This is me''' , return_tensors='''pt''' ) A_ : Tuple = model.to_bettertransformer() self.assertTrue(any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) A_ : Dict = model.generate(**_SCREAMING_SNAKE_CASE ) A_ : Union[str, Any] = model.reverse_bettertransformer() self.assertFalse(any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_SCREAMING_SNAKE_CASE ) A_ : List[Any] = AutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertFalse( any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules() ) ) A_ : str = model_reloaded.generate(**_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) def _snake_case ( self )->Optional[Any]: '''simple docstring''' A_ : List[str] = '''hf-internal-testing/tiny-random-t5''' A_ : Dict = AutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE ) A_ : List[Any] = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(_SCREAMING_SNAKE_CASE ): model.save_pretrained(_SCREAMING_SNAKE_CASE ) A_ : List[str] = model.reverse_bettertransformer() model.save_pretrained(_SCREAMING_SNAKE_CASE )
65
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Union[str, Any] = logging.get_logger(__name__) _A : List[str] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class a__ ( __lowerCAmelCase ): __lowerCAmelCase = """audio-spectrogram-transformer""" def __init__( self , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.0 , _a=0.0 , _a=0.0_2 , _a=1E-12 , _a=16 , _a=True , _a=10 , _a=10 , _a=1_024 , _a=128 , **_a , ): super().__init__(**_a ) lowercase : Any = hidden_size lowercase : int = num_hidden_layers lowercase : Union[str, Any] = num_attention_heads lowercase : Any = intermediate_size lowercase : Union[str, Any] = hidden_act lowercase : int = hidden_dropout_prob lowercase : str = attention_probs_dropout_prob lowercase : str = initializer_range lowercase : Tuple = layer_norm_eps lowercase : Any = patch_size lowercase : int = qkv_bias lowercase : Optional[Any] = frequency_stride lowercase : int = time_stride lowercase : List[str] = max_length lowercase : Tuple = num_mel_bins
202
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable def UpperCAmelCase ( lowerCamelCase_ :Callable[[int | float], int | float] , lowerCamelCase_ :int | float , lowerCamelCase_ :int | float , lowerCamelCase_ :int = 1_00 , ): '''simple docstring''' snake_case_ : Tuple = x_start snake_case_ : Optional[int] = fnc(lowerCamelCase_ ) snake_case_ : Optional[int] = 0.0 for _ in range(lowerCamelCase_ ): # Approximates small segments of curve as linear and solve # for trapezoidal area snake_case_ : int = (x_end - x_start) / steps + xa snake_case_ : Union[str, Any] = fnc(lowerCamelCase_ ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step snake_case_ : Any = xa snake_case_ : str = fxa return area if __name__ == "__main__": def UpperCAmelCase ( lowerCamelCase_ :Any ): '''simple docstring''' return x**3 + x**2 print('f(x) = x^3 + x^2') print('The area between the curve, x = -5, x = 5 and the x axis is:') __A : List[str] = 10 while i <= 100_000: print(F'with {i} steps: {trapezoidal_area(f, -5, 5, i)}') i *= 10
370
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __UpperCamelCase ( unittest.TestCase ): def __init__( self :List[Any] ,_UpperCamelCase :List[str] ,_UpperCamelCase :Optional[Any]=7 ,_UpperCamelCase :Union[str, Any]=3 ,_UpperCamelCase :Any=1_8 ,_UpperCamelCase :Optional[Any]=3_0 ,_UpperCamelCase :List[str]=4_0_0 ,_UpperCamelCase :Optional[Any]=True ,_UpperCamelCase :Union[str, Any]=None ,_UpperCamelCase :List[Any]=True ,): snake_case_ : List[str] = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : str = batch_size snake_case_ : List[Any] = num_channels snake_case_ : Tuple = image_size snake_case_ : int = min_resolution snake_case_ : int = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[Any] = size snake_case_ : Any = apply_ocr def a__ ( self :Union[str, Any] ): return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class __UpperCamelCase ( lowercase__ , unittest.TestCase ): lowercase : Tuple = LayoutLMvaImageProcessor if is_pytesseract_available() else None def a__ ( self :List[Any] ): snake_case_ : Union[str, Any] = LayoutLMvaImageProcessingTester(self ) @property def a__ ( self :int ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self :Any ): snake_case_ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCamelCase ,"""do_resize""" ) ) self.assertTrue(hasattr(_UpperCamelCase ,"""size""" ) ) self.assertTrue(hasattr(_UpperCamelCase ,"""apply_ocr""" ) ) def a__ ( self :int ): snake_case_ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 1_8, """width""": 1_8} ) snake_case_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ) self.assertEqual(image_processor.size ,{"""height""": 4_2, """width""": 4_2} ) def a__ ( self :Optional[Any] ): pass def a__ ( self :Union[str, Any] ): # Initialize image_processing snake_case_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : List[str] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase ,Image.Image ) # Test not batched input snake_case_ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ) self.assertEqual( encoding.pixel_values.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) self.assertIsInstance(encoding.words ,_UpperCamelCase ) self.assertIsInstance(encoding.boxes ,_UpperCamelCase ) # Test batched snake_case_ : List[Any] = image_processing(_UpperCamelCase ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def a__ ( self :Tuple ): # Initialize image_processing snake_case_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCamelCase ,numpify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase ,np.ndarray ) # Test not batched input snake_case_ : Optional[int] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched snake_case_ : Any = image_processing(_UpperCamelCase ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def a__ ( self :Optional[Any] ): # Initialize image_processing snake_case_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Optional[int] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCamelCase ,torchify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase ,torch.Tensor ) # Test not batched input snake_case_ : Tuple = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched snake_case_ : Union[str, Any] = image_processing(_UpperCamelCase ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def a__ ( self :List[Any] ): # with apply_OCR = True snake_case_ : Any = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case_ : List[Any] = load_dataset("""hf-internal-testing/fixtures_docvqa""" ,split="""test""" ) snake_case_ : str = Image.open(ds[0]["""file"""] ).convert("""RGB""" ) snake_case_ : Dict = image_processing(_UpperCamelCase ,return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 2_2_4, 2_2_4) ) self.assertEqual(len(encoding.words ) ,len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case_ : Tuple = [["""11:14""", """to""", """11:39""", """a.m""", """11:39""", """to""", """11:44""", """a.m.""", """11:44""", """a.m.""", """to""", """12:25""", """p.m.""", """12:25""", """to""", """12:58""", """p.m.""", """12:58""", """to""", """4:00""", """p.m.""", """2:00""", """to""", """5:00""", """p.m.""", """Coffee""", """Break""", """Coffee""", """will""", """be""", """served""", """for""", """men""", """and""", """women""", """in""", """the""", """lobby""", """adjacent""", """to""", """exhibit""", """area.""", """Please""", """move""", """into""", """exhibit""", """area.""", """(Exhibits""", """Open)""", """TRRF""", """GENERAL""", """SESSION""", """(PART""", """|)""", """Presiding:""", """Lee""", """A.""", """Waller""", """TRRF""", """Vice""", """President""", """“Introductory""", """Remarks”""", """Lee""", """A.""", """Waller,""", """TRRF""", """Vice""", """Presi-""", """dent""", """Individual""", """Interviews""", """with""", """TRRF""", """Public""", """Board""", """Members""", """and""", """Sci-""", """entific""", """Advisory""", """Council""", """Mem-""", """bers""", """Conducted""", """by""", """TRRF""", """Treasurer""", """Philip""", """G.""", """Kuehn""", """to""", """get""", """answers""", """which""", """the""", """public""", """refrigerated""", """warehousing""", """industry""", """is""", """looking""", """for.""", """Plus""", """questions""", """from""", """the""", """floor.""", """Dr.""", """Emil""", """M.""", """Mrak,""", """University""", """of""", """Cal-""", """ifornia,""", """Chairman,""", """TRRF""", """Board;""", """Sam""", """R.""", """Cecil,""", """University""", """of""", """Georgia""", """College""", """of""", """Agriculture;""", """Dr.""", """Stanley""", """Charm,""", """Tufts""", """University""", """School""", """of""", """Medicine;""", """Dr.""", """Robert""", """H.""", """Cotton,""", """ITT""", """Continental""", """Baking""", """Company;""", """Dr.""", """Owen""", """Fennema,""", """University""", """of""", """Wis-""", """consin;""", """Dr.""", """Robert""", """E.""", """Hardenburg,""", """USDA.""", """Questions""", """and""", """Answers""", """Exhibits""", """Open""", """Capt.""", """Jack""", """Stoney""", """Room""", """TRRF""", """Scientific""", """Advisory""", """Council""", """Meeting""", """Ballroom""", """Foyer"""]] # noqa: E231 snake_case_ : Any = [[[1_4_1, 5_7, 2_1_4, 6_9], [2_2_8, 5_8, 2_5_2, 6_9], [1_4_1, 7_5, 2_1_6, 8_8], [2_3_0, 7_9, 2_8_0, 8_8], [1_4_2, 2_6_0, 2_1_8, 2_7_3], [2_3_0, 2_6_1, 2_5_5, 2_7_3], [1_4_3, 2_7_9, 2_1_8, 2_9_0], [2_3_1, 2_8_2, 2_9_0, 2_9_1], [1_4_3, 3_4_2, 2_1_8, 3_5_4], [2_3_1, 3_4_5, 2_8_9, 3_5_5], [2_0_2, 3_6_2, 2_2_7, 3_7_3], [1_4_3, 3_7_9, 2_2_0, 3_9_2], [2_3_1, 3_8_2, 2_9_1, 3_9_4], [1_4_4, 7_1_4, 2_2_0, 7_2_6], [2_3_1, 7_1_5, 2_5_6, 7_2_6], [1_4_4, 7_3_2, 2_2_0, 7_4_5], [2_3_2, 7_3_6, 2_9_1, 7_4_7], [1_4_4, 7_6_9, 2_1_8, 7_8_2], [2_3_1, 7_7_0, 2_5_6, 7_8_2], [1_4_1, 7_8_8, 2_0_2, 8_0_1], [2_1_5, 7_9_1, 2_7_4, 8_0_4], [1_4_3, 8_2_6, 2_0_4, 8_3_8], [2_1_5, 8_2_6, 2_4_0, 8_3_8], [1_4_2, 8_4_4, 2_0_2, 8_5_7], [2_1_5, 8_4_7, 2_7_4, 8_5_9], [3_3_4, 5_7, 4_2_7, 6_9], [4_4_0, 5_7, 5_2_2, 6_9], [3_6_9, 7_5, 4_6_1, 8_8], [4_6_9, 7_5, 5_1_6, 8_8], [5_2_8, 7_6, 5_6_2, 8_8], [5_7_0, 7_6, 6_6_7, 8_8], [6_7_5, 7_5, 7_1_1, 8_7], [7_2_1, 7_9, 7_7_8, 8_8], [7_8_9, 7_5, 8_4_0, 8_8], [3_6_9, 9_7, 4_7_0, 1_0_7], [4_8_4, 9_4, 5_0_7, 1_0_6], [5_1_8, 9_4, 5_6_2, 1_0_7], [5_7_6, 9_4, 6_5_5, 1_1_0], [6_6_8, 9_4, 7_9_2, 1_0_9], [8_0_4, 9_5, 8_2_9, 1_0_7], [3_6_9, 1_1_3, 4_6_5, 1_2_5], [4_7_7, 1_1_6, 5_4_7, 1_2_5], [5_6_2, 1_1_3, 6_5_8, 1_2_5], [6_7_1, 1_1_6, 7_4_8, 1_2_5], [7_6_1, 1_1_3, 8_1_1, 1_2_5], [3_6_9, 1_3_1, 4_6_5, 1_4_3], [4_7_7, 1_3_3, 5_4_8, 1_4_3], [5_6_3, 1_3_0, 6_9_8, 1_4_5], [7_1_0, 1_3_0, 8_0_2, 1_4_6], [3_3_6, 1_7_1, 4_1_2, 1_8_3], [4_2_3, 1_7_1, 5_7_2, 1_8_3], [5_8_2, 1_7_0, 7_1_6, 1_8_4], [7_2_8, 1_7_1, 8_1_7, 1_8_7], [8_2_9, 1_7_1, 8_4_4, 1_8_6], [3_3_8, 1_9_7, 4_8_2, 2_1_2], [5_0_7, 1_9_6, 5_5_7, 2_0_9], [5_6_9, 1_9_6, 5_9_5, 2_0_8], [6_1_0, 1_9_6, 7_0_2, 2_0_9], [5_0_5, 2_1_4, 5_8_3, 2_2_6], [5_9_5, 2_1_4, 6_5_6, 2_2_7], [6_7_0, 2_1_5, 8_0_7, 2_2_7], [3_3_5, 2_5_9, 5_4_3, 2_7_4], [5_5_6, 2_5_9, 7_0_8, 2_7_2], [3_7_2, 2_7_9, 4_2_2, 2_9_1], [4_3_5, 2_7_9, 4_6_0, 2_9_1], [4_7_4, 2_7_9, 5_7_4, 2_9_2], [5_8_7, 2_7_8, 6_6_4, 2_9_1], [6_7_6, 2_7_8, 7_3_8, 2_9_1], [7_5_1, 2_7_9, 8_3_4, 2_9_1], [3_7_2, 2_9_8, 4_3_4, 3_1_0], [3_3_5, 3_4_1, 4_8_3, 3_5_4], [4_9_7, 3_4_1, 6_5_5, 3_5_4], [6_6_7, 3_4_1, 7_2_8, 3_5_4], [7_4_0, 3_4_1, 8_2_5, 3_5_4], [3_3_5, 3_6_0, 4_3_0, 3_7_2], [4_4_2, 3_6_0, 5_3_4, 3_7_2], [5_4_5, 3_5_9, 6_8_7, 3_7_2], [6_9_7, 3_6_0, 7_5_4, 3_7_2], [7_6_5, 3_6_0, 8_2_3, 3_7_3], [3_3_4, 3_7_8, 4_2_8, 3_9_1], [4_4_0, 3_7_8, 5_7_7, 3_9_4], [5_9_0, 3_7_8, 7_0_5, 3_9_1], [7_2_0, 3_7_8, 8_0_1, 3_9_1], [3_3_4, 3_9_7, 4_0_0, 4_0_9], [3_7_0, 4_1_6, 5_2_9, 4_2_9], [5_4_4, 4_1_6, 5_7_6, 4_3_2], [5_8_7, 4_1_6, 6_6_5, 4_2_8], [6_7_7, 4_1_6, 8_1_4, 4_2_9], [3_7_2, 4_3_5, 4_5_2, 4_5_0], [4_6_5, 4_3_4, 4_9_5, 4_4_7], [5_1_1, 4_3_4, 6_0_0, 4_4_7], [6_1_1, 4_3_6, 6_3_7, 4_4_7], [6_4_9, 4_3_6, 6_9_4, 4_5_1], [7_0_5, 4_3_8, 8_2_4, 4_4_7], [3_6_9, 4_5_3, 4_5_2, 4_6_6], [4_6_4, 4_5_4, 5_0_9, 4_6_6], [5_2_2, 4_5_3, 6_1_1, 4_6_9], [6_2_5, 4_5_3, 7_9_2, 4_6_9], [3_7_0, 4_7_2, 5_5_6, 4_8_8], [5_7_0, 4_7_2, 6_8_4, 4_8_7], [6_9_7, 4_7_2, 7_1_8, 4_8_5], [7_3_2, 4_7_2, 8_3_5, 4_8_8], [3_6_9, 4_9_0, 4_1_1, 5_0_3], [4_2_5, 4_9_0, 4_8_4, 5_0_3], [4_9_6, 4_9_0, 6_3_5, 5_0_6], [6_4_5, 4_9_0, 7_0_7, 5_0_3], [7_1_8, 4_9_1, 7_6_1, 5_0_3], [7_7_1, 4_9_0, 8_4_0, 5_0_3], [3_3_6, 5_1_0, 3_7_4, 5_2_1], [3_8_8, 5_1_0, 4_4_7, 5_2_2], [4_6_0, 5_1_0, 4_8_9, 5_2_1], [5_0_3, 5_1_0, 5_8_0, 5_2_2], [5_9_2, 5_0_9, 7_3_6, 5_2_5], [7_4_5, 5_0_9, 7_7_0, 5_2_2], [7_8_1, 5_0_9, 8_4_0, 5_2_2], [3_3_8, 5_2_8, 4_3_4, 5_4_1], [4_4_8, 5_2_8, 5_9_6, 5_4_1], [6_0_9, 5_2_7, 6_8_7, 5_4_0], [7_0_0, 5_2_8, 7_9_2, 5_4_1], [3_3_6, 5_4_6, 3_9_7, 5_5_9], [4_0_7, 5_4_6, 4_3_1, 5_5_9], [4_4_3, 5_4_6, 5_2_5, 5_6_0], [5_3_7, 5_4_6, 6_8_0, 5_6_2], [6_8_8, 5_4_6, 7_1_4, 5_5_9], [7_2_2, 5_4_6, 8_3_7, 5_6_2], [3_3_6, 5_6_5, 4_4_9, 5_8_1], [4_6_1, 5_6_5, 4_8_5, 5_7_7], [4_9_7, 5_6_5, 6_6_5, 5_8_1], [6_8_1, 5_6_5, 7_1_8, 5_7_7], [7_3_2, 5_6_5, 8_3_7, 5_8_0], [3_3_7, 5_8_4, 4_3_8, 5_9_7], [4_5_2, 5_8_3, 5_2_1, 5_9_6], [5_3_5, 5_8_4, 6_7_7, 5_9_9], [6_9_0, 5_8_3, 7_8_7, 5_9_6], [8_0_1, 5_8_3, 8_2_5, 5_9_6], [3_3_8, 6_0_2, 4_7_8, 6_1_5], [4_9_2, 6_0_2, 5_3_0, 6_1_4], [5_4_3, 6_0_2, 6_3_8, 6_1_5], [6_5_0, 6_0_2, 6_7_6, 6_1_4], [6_8_8, 6_0_2, 7_8_8, 6_1_5], [8_0_2, 6_0_2, 8_4_3, 6_1_4], [3_3_7, 6_2_1, 5_0_2, 6_3_3], [5_1_6, 6_2_1, 6_1_5, 6_3_7], [6_2_9, 6_2_1, 7_7_4, 6_3_6], [7_8_9, 6_2_1, 8_2_7, 6_3_3], [3_3_7, 6_3_9, 4_1_8, 6_5_2], [4_3_2, 6_4_0, 5_7_1, 6_5_3], [5_8_7, 6_3_9, 7_3_1, 6_5_5], [7_4_3, 6_3_9, 7_6_9, 6_5_2], [7_8_0, 6_3_9, 8_4_1, 6_5_2], [3_3_8, 6_5_8, 4_4_0, 6_7_3], [4_5_5, 6_5_8, 4_9_1, 6_7_0], [5_0_8, 6_5_8, 6_0_2, 6_7_1], [6_1_6, 6_5_8, 6_3_8, 6_7_0], [6_5_4, 6_5_8, 8_3_5, 6_7_4], [3_3_7, 6_7_7, 4_2_9, 6_8_9], [3_3_7, 7_1_4, 4_8_2, 7_2_6], [4_9_5, 7_1_4, 5_4_8, 7_2_6], [5_6_1, 7_1_4, 6_8_3, 7_2_6], [3_3_8, 7_7_0, 4_6_1, 7_8_2], [4_7_4, 7_6_9, 5_5_4, 7_8_5], [4_8_9, 7_8_8, 5_6_2, 8_0_3], [5_7_6, 7_8_8, 6_4_3, 8_0_1], [6_5_6, 7_8_7, 7_5_1, 8_0_4], [7_6_4, 7_8_8, 8_4_4, 8_0_1], [3_3_4, 8_2_5, 4_2_1, 8_3_8], [4_3_0, 8_2_4, 5_7_4, 8_3_8], [5_8_4, 8_2_4, 7_2_3, 8_4_1], [3_3_5, 8_4_4, 4_5_0, 8_5_7], [4_6_4, 8_4_3, 5_8_3, 8_6_0], [6_2_8, 8_6_2, 7_5_5, 8_7_5], [7_6_9, 8_6_1, 8_4_8, 8_7_8]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words ,_UpperCamelCase ) self.assertListEqual(encoding.boxes ,_UpperCamelCase ) # with apply_OCR = False snake_case_ : Dict = LayoutLMvaImageProcessor(apply_ocr=_UpperCamelCase ) snake_case_ : Optional[int] = image_processing(_UpperCamelCase ,return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 2_2_4, 2_2_4) )
8
0