code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return numa ^ numa < 0 if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = CustomTokenizer pass
286
1
"""simple docstring""" lowerCamelCase_ : List[str] = tuple[float, float, float] lowerCamelCase_ : Optional[int] = tuple[float, float, float] def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = end_pointa[0] - end_pointa[0] A_ : Dict = end_pointa[1] - end_pointa[1] A_ : List[Any] = end_pointa[2] - end_pointa[2] return (x, y, z) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[str] = ab[1] * ac[2] - ab[2] * ac[1] # *i A_ : int = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j A_ : Optional[Any] = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return tuple(round(_UpperCAmelCase , _UpperCAmelCase ) for x in vector ) == (0, 0, 0) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 10 ): """simple docstring""" A_ : Optional[int] = create_vector(_UpperCAmelCase , _UpperCAmelCase ) A_ : Dict = create_vector(_UpperCAmelCase , _UpperCAmelCase ) return is_zero_vector(get_ad_vectors_cross(_UpperCAmelCase , _UpperCAmelCase ) , _UpperCAmelCase )
286
"""simple docstring""" from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch lowerCamelCase_ : str = logging.get_logger(__name__) @add_end_docstrings( UpperCAmelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.framework == "tf": A_ : str = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": A_ : List[str] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ) else: raise ValueError('Unsupported framework' ) return masked_index def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = self.get_masked_index(snake_case_ ) A_ : str = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['input_ids'][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_=None , **snake_case_ ): """simple docstring""" if return_tensors is None: A_ : Any = self.framework A_ : Dict = self.tokenizer(snake_case_ , return_tensors=snake_case_ ) self.ensure_exactly_one_mask_token(snake_case_ ) return model_inputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Dict = self.model(**snake_case_ ) A_ : Optional[int] = model_inputs['input_ids'] return model_outputs def lowerCamelCase_ ( self , snake_case_ , snake_case_=5 , snake_case_=None ): """simple docstring""" if target_ids is not None and target_ids.shape[0] < top_k: A_ : str = target_ids.shape[0] A_ : Optional[Any] = model_outputs['input_ids'][0] A_ : List[Any] = model_outputs['logits'] if self.framework == "tf": A_ : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] A_ : Union[str, Any] = outputs.numpy() A_ : Optional[int] = outputs[0, masked_index, :] A_ : Optional[Any] = stable_softmax(snake_case_ , axis=-1 ) if target_ids is not None: A_ : Union[str, Any] = tf.gather_nd(tf.squeeze(snake_case_ , 0 ) , target_ids.reshape(-1 , 1 ) ) A_ : Optional[int] = tf.expand_dims(snake_case_ , 0 ) A_ : Any = tf.math.top_k(snake_case_ , k=snake_case_ ) A_ , A_ : str = topk.values.numpy(), topk.indices.numpy() else: A_ : int = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample A_ : Tuple = outputs[0, masked_index, :] A_ : List[str] = logits.softmax(dim=-1 ) if target_ids is not None: A_ : str = probs[..., target_ids] A_ , A_ : List[str] = probs.topk(snake_case_ ) A_ : List[Any] = [] A_ : int = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): A_ : str = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place A_ : Union[str, Any] = input_ids.numpy().copy() if target_ids is not None: A_ : str = target_ids[p].tolist() A_ : Union[str, Any] = p # Filter padding out: A_ : Any = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back A_ : Any = self.tokenizer.decode(snake_case_ , skip_special_tokens=snake_case_ ) A_ : Any = {'score': v, 'token': p, 'token_str': self.tokenizer.decode([p] ), 'sequence': sequence} row.append(snake_case_ ) result.append(snake_case_ ) if single_mask: return result[0] return result def lowerCamelCase_ ( self , snake_case_ , snake_case_=None ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : List[str] = [targets] try: A_ : Optional[int] = self.tokenizer.get_vocab() except Exception: A_ : int = {} A_ : Tuple = [] for target in targets: A_ : int = vocab.get(snake_case_ , snake_case_ ) if id_ is None: A_ : Tuple = self.tokenizer( snake_case_ , add_special_tokens=snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , max_length=1 , truncation=snake_case_ , )['input_ids'] if len(snake_case_ ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ 'We cannot replace it with anything meaningful, ignoring it' ) continue A_ : str = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) A_ : Tuple = list(set(snake_case_ ) ) if len(snake_case_ ) == 0: raise ValueError('At least one target must be provided when passed.' ) A_ : Optional[Any] = np.array(snake_case_ ) return target_ids def lowerCamelCase_ ( self , snake_case_=None , snake_case_=None ): """simple docstring""" A_ : List[str] = {} if targets is not None: A_ : Any = self.get_target_ids(snake_case_ , snake_case_ ) A_ : Optional[Any] = target_ids if top_k is not None: A_ : int = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , 'The tokenizer does not define a `mask_token`.' ) return {}, {}, postprocess_params def __call__( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : List[str] = super().__call__(snake_case_ , **snake_case_ ) if isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1: return outputs[0] return outputs
286
1
"""simple docstring""" import os import pytest from attr import dataclass lowerCamelCase_ : List[Any] = 'us-east-1' # defaults region @dataclass class _UpperCAmelCase : '''simple docstring''' lowercase_ : str lowercase_ : Tuple = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" lowercase_ : List[str] = { """task_name""": """mnli""", """per_device_train_batch_size""": 16, """per_device_eval_batch_size""": 16, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 500, """save_steps""": 5_500, } lowercase_ : Tuple = {**hyperparameters, """max_steps""": 1_000} @property def lowerCamelCase_ ( self ): """simple docstring""" if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def lowerCamelCase_ ( self ): """simple docstring""" return F"""{self.framework}-transfromers-test""" @property def lowerCamelCase_ ( self ): """simple docstring""" return F"""./tests/sagemaker/scripts/{self.framework}""" @property def lowerCamelCase_ ( self ): """simple docstring""" if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = SageMakerTestEnvironment(framework=request.cls.framework )
286
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
1
"""simple docstring""" import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = tempfile.mkdtemp() A_ : str = 5 # Realm tok A_ : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'test', 'question', 'this', 'is', 'the', 'first', 'second', 'third', 'fourth', 'fifth', 'record', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] A_ : Union[str, Any] = os.path.join(self.tmpdirname , 'realm_tokenizer' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) A_ : int = os.path.join(snake_case_ , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) A_ : List[str] = os.path.join(self.tmpdirname , 'realm_block_records' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'realm_tokenizer' ) ) def lowerCamelCase_ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[int] = RealmConfig(num_block_records=self.num_block_records ) return config def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = Dataset.from_dict( { 'id': ['0', '1'], 'question': ['foo', 'bar'], 'answers': [['Foo', 'Bar'], ['Bar']], } ) return dataset def lowerCamelCase_ ( self ): """simple docstring""" A_ : str = np.array( [ B'This is the first record', B'This is the second record', B'This is the third record', B'This is the fourth record', B'This is the fifth record', B'This is a longer longer longer record', ] , dtype=snake_case_ , ) return block_records def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[Any] = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = self.get_config() A_ : Tuple = self.get_dummy_retriever() A_ : Optional[Any] = retriever.tokenizer A_ : Dict = np.array([0, 3] , dtype='long' ) A_ : int = tokenizer(['Test question'] ).input_ids A_ : Dict = tokenizer( ['the fourth'] , add_special_tokens=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , ).input_ids A_ : Tuple = config.reader_seq_len A_ , A_ , A_ , A_ : int = retriever( snake_case_ , snake_case_ , answer_ids=snake_case_ , max_length=snake_case_ , return_tensors='np' ) self.assertEqual(len(snake_case_ ) , 2 ) self.assertEqual(len(snake_case_ ) , 2 ) self.assertEqual(len(snake_case_ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 1_0) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 1_0) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 1_0) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 1_0) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['[CLS]', 'test', 'question', '[SEP]', 'this', 'is', 'the', 'first', 'record', '[SEP]'] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['[CLS]', 'test', 'question', '[SEP]', 'this', 'is', 'the', 'fourth', 'record', '[SEP]'] , ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : str = self.get_config() A_ : int = self.get_dummy_retriever() A_ : Dict = retriever.tokenizer A_ : int = np.array([0, 3, 5] , dtype='long' ) A_ : Union[str, Any] = tokenizer(['Test question'] ).input_ids A_ : str = tokenizer( ['the fourth', 'longer longer'] , add_special_tokens=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , ).input_ids A_ : Any = config.reader_seq_len A_ , A_ , A_ , A_ : str = retriever( snake_case_ , snake_case_ , answer_ids=snake_case_ , max_length=snake_case_ , return_tensors='np' ) self.assertEqual([False, True, True] , snake_case_ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , snake_case_ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , 'realm_block_records' ) ) # Test local path A_ : str = retriever.from_pretrained(os.path.join(self.tmpdirname , 'realm_block_records' ) ) self.assertEqual(retriever.block_records[0] , B'This is the first record' ) # Test mocked remote path with patch('transformers.models.realm.retrieval_realm.hf_hub_download' ) as mock_hf_hub_download: A_ : Any = os.path.join( os.path.join(self.tmpdirname , 'realm_block_records' ) , _REALM_BLOCK_RECORDS_FILENAME ) A_ : Union[str, Any] = RealmRetriever.from_pretrained('google/realm-cc-news-pretrained-openqa' ) self.assertEqual(retriever.block_records[0] , B'This is the first record' )
286
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) A_ : Any = hex_num[0] == '-' if is_negative: A_ : Optional[Any] = hex_num[1:] try: A_ : Tuple = int(_UpperCAmelCase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) A_ : Union[str, Any] = '' while int_num > 0: A_ : Optional[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" # limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( 'pipelines_utils', '0.22.0', 'Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.', standard_warn=False, stacklevel=3, )
286
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
1
"""simple docstring""" import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ , snake_case_=1_3 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=9_9 , snake_case_=3_2 , snake_case_=5 , snake_case_=4 , snake_case_=3_7 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_0 , snake_case_=0.02 , snake_case_=True , snake_case_=None , ): """simple docstring""" A_ : List[Any] = parent A_ : Dict = batch_size A_ : Union[str, Any] = seq_length A_ : Union[str, Any] = is_training A_ : int = use_input_mask A_ : Dict = vocab_size A_ : Tuple = hidden_size A_ : List[str] = num_hidden_layers A_ : Any = num_attention_heads A_ : Optional[int] = intermediate_size A_ : Optional[int] = hidden_act A_ : int = hidden_dropout_prob A_ : str = attention_probs_dropout_prob A_ : Union[str, Any] = max_position_embeddings A_ : List[str] = initializer_range A_ : str = use_labels A_ : Optional[int] = scope def lowerCamelCase_ ( self ): """simple docstring""" A_ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Any = None if self.use_input_mask: A_ : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: A_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : str = self.get_config() return config, input_ids, input_mask, token_labels def lowerCamelCase_ ( self ): """simple docstring""" return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=snake_case_ , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self ): """simple docstring""" ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) : Optional[Any] = self.prepare_config_and_inputs() A_ : Any = True A_ : Dict = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ , ): """simple docstring""" A_ : Dict = BertGenerationEncoder(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : int = model(snake_case_ , attention_mask=snake_case_ ) A_ : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ , ): """simple docstring""" A_ : Optional[int] = True A_ : List[str] = BertGenerationEncoder(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : int = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , ) A_ : int = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ , ): """simple docstring""" A_ : str = True A_ : List[Any] = True A_ : Any = BertGenerationDecoder(config=snake_case_ ).to(snake_case_ ).eval() # first forward pass A_ : Optional[Any] = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , use_cache=snake_case_ , ) A_ : Optional[int] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A_ : Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A_ : str = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A_ : List[str] = torch.cat([input_ids, next_tokens] , dim=-1 ) A_ : List[Any] = torch.cat([input_mask, next_mask] , dim=-1 ) A_ : Union[str, Any] = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , output_hidden_states=snake_case_ , )['hidden_states'][0] A_ : Optional[Any] = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , past_key_values=snake_case_ , output_hidden_states=snake_case_ , )['hidden_states'][0] # select random slice A_ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A_ : int = output_from_no_past[:, -3:, random_slice_idx].detach() A_ : int = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case_ , snake_case_ , atol=1E-3 ) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): """simple docstring""" A_ : str = BertGenerationDecoder(snake_case_ ) model.to(snake_case_ ) model.eval() A_ : Dict = model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ , A_ , A_ : Union[str, Any] = self.prepare_config_and_inputs() A_ : str = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowercase_ : List[str] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () lowercase_ : str = (BertGenerationDecoder,) if is_torch_available() else () lowercase_ : Any = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = BertGenerationEncoderTester(self ) A_ : int = ConfigTester(self , config_class=snake_case_ , hidden_size=3_7 ) def lowerCamelCase_ ( self ): """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase_ ( self ): """simple docstring""" A_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ , A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs() A_ : Any = 'bert' self.model_tester.create_and_check_model(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) : str = self.model_tester.prepare_config_and_inputs_for_decoder() A_ : str = None self.model_tester.create_and_check_model_as_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*snake_case_ ) @slow def lowerCamelCase_ ( self ): """simple docstring""" A_ : Tuple = BertGenerationEncoder.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) self.assertIsNotNone(snake_case_ ) @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[int] = BertGenerationEncoder.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) A_ : Tuple = torch.tensor([[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_2_6, 3_8_9_9, 2_0_0_3, 1_0_1_4_0, 1_0_2]] ) with torch.no_grad(): A_ : Any = model(snake_case_ )[0] A_ : str = torch.Size([1, 8, 1_0_2_4] ) self.assertEqual(output.shape , snake_case_ ) A_ : Dict = torch.tensor( [[[0.17_75, 0.00_83, -0.03_21], [1.60_02, 0.12_87, 0.39_12], [2.14_73, 0.57_91, 0.60_66]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case_ , atol=1E-4 ) ) @require_torch class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = BertGenerationDecoder.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) A_ : Any = torch.tensor([[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_2_6, 3_8_9_9, 2_0_0_3, 1_0_1_4_0, 1_0_2]] ) with torch.no_grad(): A_ : Any = model(snake_case_ )[0] A_ : int = torch.Size([1, 8, 5_0_3_5_8] ) self.assertEqual(output.shape , snake_case_ ) A_ : int = torch.tensor( [[[-0.57_88, -2.59_94, -3.70_54], [0.04_38, 4.79_97, 1.87_95], [1.58_62, 6.64_09, 4.46_38]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case_ , atol=1E-4 ) )
286
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Any = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """xmod""" def __init__( self , snake_case_=3_0_5_2_2 , snake_case_=7_6_8 , snake_case_=1_2 , snake_case_=1_2 , snake_case_=3_0_7_2 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , snake_case_=False , snake_case_=2 , snake_case_=False , snake_case_=True , snake_case_=True , snake_case_=("en_XX",) , snake_case_=None , **snake_case_ , ): """simple docstring""" super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) A_ : Union[str, Any] = vocab_size A_ : Any = hidden_size A_ : List[str] = num_hidden_layers A_ : Tuple = num_attention_heads A_ : int = hidden_act A_ : Any = intermediate_size A_ : Any = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Union[str, Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : List[str] = initializer_range A_ : Any = layer_norm_eps A_ : Optional[Any] = position_embedding_type A_ : int = use_cache A_ : Dict = classifier_dropout A_ : int = pre_norm A_ : Optional[Any] = adapter_reduction_factor A_ : List[Any] = adapter_layer_norm A_ : int = adapter_reuse_layer_norm A_ : Dict = ln_before_adapter A_ : List[str] = list(snake_case_ ) A_ : Union[str, Any] = default_language class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @property def lowerCamelCase_ ( self ): """simple docstring""" if self.task == "multiple-choice": A_ : Dict = {0: 'batch', 1: 'choice', 2: 'sequence'} else: A_ : int = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return abs(_UpperCAmelCase ) if a == 0 else greatest_common_divisor(b % a , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" while y: # --> when y=0 then loop will terminate and return x as final GCD. A_ , A_ : Tuple = y, x % y return abs(_UpperCAmelCase ) def UpperCAmelCase__ ( ): """simple docstring""" try: A_ : List[str] = input('Enter two integers separated by comma (,): ' ).split(',' ) A_ : List[str] = int(nums[0] ) A_ : List[Any] = int(nums[1] ) print( f"""greatest_common_divisor({num_a}, {num_a}) = """ f"""{greatest_common_divisor(_UpperCAmelCase , _UpperCAmelCase )}""" ) print(f"""By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(_UpperCAmelCase , _UpperCAmelCase )}""" ) except (IndexError, UnboundLocalError, ValueError): print('Wrong input' ) if __name__ == "__main__": main()
286
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ : Optional[int] = logging.get_logger(__name__) lowerCamelCase_ : List[Any] = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = """switch_transformers""" lowercase_ : str = ["""past_key_values"""] lowercase_ : Dict = {"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , snake_case_=3_2_1_2_8 , snake_case_=7_6_8 , snake_case_=6_4 , snake_case_=2_0_4_8 , snake_case_=6_4 , snake_case_=1_2 , snake_case_=3 , snake_case_=1_2 , snake_case_=3 , snake_case_=1_2 , snake_case_=8 , snake_case_=False , snake_case_=0.01 , snake_case_="float32" , snake_case_=False , snake_case_=3_2 , snake_case_=1_2_8 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=0.0_01 , snake_case_=0.0_01 , snake_case_=1.0 , snake_case_="relu" , snake_case_=True , snake_case_=False , snake_case_=True , snake_case_=0 , snake_case_=1 , **snake_case_ , ): """simple docstring""" A_ : str = vocab_size A_ : Tuple = d_model A_ : Tuple = d_kv A_ : str = d_ff A_ : Tuple = num_sparse_encoder_layers A_ : int = num_layers A_ : List[Any] = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry A_ : List[str] = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: A_ : List[str] = self.num_layers // self.num_sparse_encoder_layers else: A_ : str = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: A_ : int = self.num_decoder_layers // self.num_sparse_decoder_layers else: A_ : str = self.num_decoder_layers # HACK: this will create 0 sparse layers A_ : Optional[Any] = num_heads A_ : Tuple = num_experts A_ : List[str] = expert_capacity A_ : List[Any] = router_bias A_ : Optional[int] = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) A_ : Dict = router_dtype A_ : int = router_ignore_padding_tokens A_ : Union[str, Any] = relative_attention_num_buckets A_ : List[str] = relative_attention_max_distance A_ : str = dropout_rate A_ : Tuple = layer_norm_epsilon A_ : Optional[Any] = initializer_factor A_ : str = feed_forward_proj A_ : List[Any] = use_cache A_ : Optional[Any] = add_router_probs A_ : Optional[Any] = router_z_loss_coef A_ : Union[str, Any] = router_aux_loss_coef A_ : str = self.feed_forward_proj.split('-' ) A_ : List[Any] = act_info[-1] A_ : Tuple = act_info[0] == 'gated' if len(snake_case_ ) > 1 and act_info[0] != "gated" or len(snake_case_ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": A_ : List[Any] = 'gelu_new' super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ , )
286
"""simple docstring""" from copy import deepcopy class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None , snake_case_ = None ): """simple docstring""" if arr is None and size is not None: A_ : Union[str, Any] = size A_ : List[str] = [0] * size elif arr is not None: self.init(snake_case_ ) else: raise ValueError('Either arr or size must be specified' ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = len(snake_case_ ) A_ : Optional[int] = deepcopy(snake_case_ ) for i in range(1 , self.size ): A_ : Optional[Any] = self.next_(snake_case_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): A_ : Optional[int] = self.next_(snake_case_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index + (index & (-index)) @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index - (index & (-index)) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value A_ : List[str] = self.next_(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" self.add(snake_case_ , value - self.get(snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if right == 0: return 0 A_ : Any = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] A_ : Tuple = self.prev(snake_case_ ) return result def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" return self.prefix(snake_case_ ) - self.prefix(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.query(snake_case_ , index + 1 ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" value -= self.tree[0] if value < 0: return -1 A_ : List[Any] = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 A_ : Tuple = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging lowerCamelCase_ : Optional[int] = '\\n\n' lowerCamelCase_ : List[str] = '\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n' lowerCamelCase_ : Union[str, Any] = '\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to \'cuda\' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric("perplexity")\n >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"]\n >>> results = perplexity.compute(model_id=\'gpt2\',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n [\'perplexities\', \'mean_perplexity\']\n >>> print(round(results["mean_perplexity"], 2))\n 78.22\n >>> print(round(results["perplexities"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric("perplexity")\n >>> input_texts = datasets.load_dataset("wikitext",\n ... "wikitext-2-raw-v1",\n ... split="test")["text"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!=\'\']\n >>> results = perplexity.compute(model_id=\'gpt2\',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n [\'perplexities\', \'mean_perplexity\']\n >>> print(round(results["mean_perplexity"], 2))\n 60.35\n >>> print(round(results["perplexities"][0], 2))\n 81.12\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def lowerCamelCase_ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ = 1_6 , snake_case_ = True , snake_case_=None ): """simple docstring""" if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": A_ : Optional[int] = 'cuda' else: A_ : List[Any] = 'cuda' if torch.cuda.is_available() else 'cpu' A_ : str = AutoModelForCausalLM.from_pretrained(snake_case_ ) A_ : Any = model.to(snake_case_ ) A_ : Any = AutoTokenizer.from_pretrained(snake_case_ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: A_ : Optional[int] = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(snake_case_ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" A_ : str = model.config.max_length - 1 else: A_ : List[str] = model.config.max_length A_ : Dict = tokenizer( snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , return_tensors='pt' , return_attention_mask=snake_case_ , ).to(snake_case_ ) A_ : Tuple = encodings['input_ids'] A_ : Tuple = encodings['attention_mask'] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." A_ : List[Any] = [] A_ : Optional[int] = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(snake_case_ ) , snake_case_ ) ): A_ : Optional[Any] = min(start_index + batch_size , len(snake_case_ ) ) A_ : Dict = encoded_texts[start_index:end_index] A_ : Any = attn_masks[start_index:end_index] if add_start_token: A_ : int = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(snake_case_ ) A_ : Dict = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) A_ : Any = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(snake_case_ ), attn_mask] , dim=1 ) A_ : Union[str, Any] = encoded_batch with torch.no_grad(): A_ : List[Any] = model(snake_case_ , attention_mask=snake_case_ ).logits A_ : Optional[int] = out_logits[..., :-1, :].contiguous() A_ : Tuple = labels[..., 1:].contiguous() A_ : List[str] = attn_mask[..., 1:].contiguous() A_ : List[str] = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , snake_case_ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(snake_case_ )}
286
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , snake_case_ = 7_6_8 , ): """simple docstring""" super().__init__() A_ : Optional[int] = nn.Parameter(torch.zeros(1 , snake_case_ ) ) A_ : Optional[int] = nn.Parameter(torch.ones(1 , snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ = None , snake_case_ = None , ): """simple docstring""" A_ : str = nn.Parameter(self.mean.to(snake_case_ ).to(snake_case_ ) ) A_ : Optional[int] = nn.Parameter(self.std.to(snake_case_ ).to(snake_case_ ) ) return self def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Tuple = (embeds - self.mean) * 1.0 / self.std return embeds def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = (embeds * self.std) + self.mean return embeds
286
1
"""simple docstring""" from queue import PriorityQueue from typing import Any import numpy as np def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" for nxt, d in graph[v]: if nxt in visited_forward: continue A_ : Optional[int] = cst_fwd.get(_UpperCAmelCase , np.inf ) A_ : Any = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) A_ : Optional[Any] = new_cost_f A_ : str = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: A_ : Any = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = -1 A_ : Optional[Any] = set() A_ : Any = set() A_ : Any = {source: 0} A_ : Optional[int] = {destination: 0} A_ : Dict = {source: None} A_ : Optional[int] = {destination: None} A_ : PriorityQueue[Any] = PriorityQueue() A_ : PriorityQueue[Any] = PriorityQueue() A_ : Optional[Any] = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): A_ , A_ : Tuple = queue_forward.get() visited_forward.add(_UpperCAmelCase ) A_ , A_ : Any = queue_backward.get() visited_backward.add(_UpperCAmelCase ) A_ : Optional[Any] = pass_and_relaxation( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) A_ : Optional[Any] = pass_and_relaxation( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: A_ : List[str] = shortest_distance return shortest_path_distance lowerCamelCase_ : int = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } lowerCamelCase_ : List[str] = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCamelCase_ : Any = HfArgumentParser(InitializationArguments) lowerCamelCase_ : Union[str, Any] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowerCamelCase_ : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowerCamelCase_ : Tuple = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) lowerCamelCase_ : int = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowerCamelCase_ : Any = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
286
1
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm lowerCamelCase_ : List[Any] = logging.get_logger(__name__) @dataclass class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Optional[int] = [ """no_inference""", """no_cuda""", """no_tpu""", """no_speed""", """no_memory""", """no_env_print""", """no_multi_process""", ] def __init__( self , **snake_case_ ): """simple docstring""" for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: A_ : Dict = deprecated_arg[3:] setattr(self , snake_case_ , not kwargs.pop(snake_case_ ) ) logger.warning( F"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" F""" {positive_arg}={kwargs[positive_arg]}""" ) A_ : Optional[int] = kwargs.pop('torchscript' , self.torchscript ) A_ : List[Any] = kwargs.pop('torch_xla_tpu_print_metrics' , self.torch_xla_tpu_print_metrics ) A_ : Optional[int] = kwargs.pop('fp16_opt_level' , self.fpaa_opt_level ) super().__init__(**snake_case_ ) lowercase_ : bool = field(default=UpperCAmelCase__ , metadata={"""help""": """Trace the models using torchscript"""} ) lowercase_ : bool = field(default=UpperCAmelCase__ , metadata={"""help""": """Print Xla/PyTorch tpu metrics"""} ) lowercase_ : str = field( default="""O1""" , metadata={ """help""": ( """For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. """ """See details at https://nvidia.github.io/apex/amp.html""" ) } , ) @cached_property def lowerCamelCase_ ( self ): """simple docstring""" requires_backends(self , ['torch'] ) logger.info('PyTorch: setting up devices' ) if not self.cuda: A_ : Optional[Any] = torch.device('cpu' ) A_ : Optional[int] = 0 elif is_torch_tpu_available(): A_ : int = xm.xla_device() A_ : Union[str, Any] = 0 else: A_ : Dict = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) A_ : Optional[Any] = torch.cuda.device_count() return device, n_gpu @property def lowerCamelCase_ ( self ): """simple docstring""" return is_torch_tpu_available() and self.tpu @property def lowerCamelCase_ ( self ): """simple docstring""" requires_backends(self , ['torch'] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def lowerCamelCase_ ( self ): """simple docstring""" requires_backends(self , ['torch'] ) return self._setup_devices[0] @property def lowerCamelCase_ ( self ): """simple docstring""" requires_backends(self , ['torch'] ) return self._setup_devices[1] @property def lowerCamelCase_ ( self ): """simple docstring""" return self.n_gpu > 0
286
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser lowerCamelCase_ : Any = re.compile(r'\s+') def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return {"hash": hashlib.mda(re.sub(_UpperCAmelCase , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [len(_UpperCAmelCase ) for line in example['content'].splitlines()] return {"line_mean": np.mean(_UpperCAmelCase ), "line_max": max(_UpperCAmelCase )} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 ): """simple docstring""" A_ : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] A_ : List[str] = example['content'].splitlines() for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 , _UpperCAmelCase=0.05 ): """simple docstring""" A_ : Any = ['unit tests', 'test file', 'configuration file'] A_ : Dict = example['content'].splitlines() A_ : List[Any] = 0 A_ : str = 0 # first test for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test A_ : Tuple = example['content'].count('\n' ) A_ : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = ['def ', 'class ', 'for ', 'while '] A_ : Tuple = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=4 ): """simple docstring""" A_ : Union[str, Any] = example['content'].splitlines() A_ : Any = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = tokenizer(example['content'] , truncation=_UpperCAmelCase )['input_ids'] A_ : Dict = len(example['content'] ) / len(_UpperCAmelCase ) return {"ratio": ratio} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = {} results.update(get_hash(_UpperCAmelCase ) ) results.update(line_stats(_UpperCAmelCase ) ) results.update(alpha_stats(_UpperCAmelCase ) ) results.update(char_token_ratio(_UpperCAmelCase ) ) results.update(is_autogenerated(_UpperCAmelCase ) ) results.update(is_config_or_test(_UpperCAmelCase ) ) results.update(has_no_keywords(_UpperCAmelCase ) ) results.update(has_few_assignments(_UpperCAmelCase ) ) return results def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if not check_uniques(_UpperCAmelCase , _UpperCAmelCase ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'rb' ) as f_in: with gzip.open(str(_UpperCAmelCase ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) os.unlink(_UpperCAmelCase ) # Settings lowerCamelCase_ : Optional[int] = HfArgumentParser(PreprocessingArguments) lowerCamelCase_ : Optional[Any] = parser.parse_args() if args.num_workers is None: lowerCamelCase_ : int = multiprocessing.cpu_count() lowerCamelCase_ : Tuple = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset lowerCamelCase_ : Tuple = time.time() lowerCamelCase_ : Tuple = load_dataset(args.dataset_name, split='train') print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing lowerCamelCase_ : List[str] = time.time() lowerCamelCase_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes lowerCamelCase_ : int = set(ds.unique('hash')) lowerCamelCase_ : Union[str, Any] = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics lowerCamelCase_ : Optional[int] = time.time() lowerCamelCase_ : Tuple = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: lowerCamelCase_ : Union[str, Any] = time.time() lowerCamelCase_ , lowerCamelCase_ : str = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file lowerCamelCase_ : Tuple = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) lowerCamelCase_ : Optional[Any] = output_dir / 'data' data_dir.mkdir(exist_ok=True) lowerCamelCase_ : List[str] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): lowerCamelCase_ : Optional[int] = str(data_dir / F"file-{file_number+1:012}.json") lowerCamelCase_ : List[str] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
286
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor lowerCamelCase_ : Optional[int] = logging.get_logger(__name__) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" warnings.warn( 'The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use YolosImageProcessor instead.' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
286
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : Optional[Any] = [] # fmt: off # stem: rename_keys.append(('cls_token', 'vit.embeddings.cls_token') ) rename_keys.append(('pos_embed', 'vit.embeddings.position_embeddings') ) rename_keys.append(('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias') ) # backbone rename_keys.append(('patch_embed.backbone.stem.conv.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.bias', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A_ : List[str] = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) # fmt: on return rename_keys def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: A_ : List[str] = '' else: A_ : Dict = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ : str = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) A_ : List[Any] = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A_ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A_ : Tuple = in_proj_bias[: config.hidden_size] A_ : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A_ : Tuple = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = dct.pop(_UpperCAmelCase ) A_ : Optional[int] = val def UpperCAmelCase__ ( ): """simple docstring""" A_ : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' A_ : int = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : List[Any] = BitConfig( global_padding='same' , layer_type='bottleneck' , depths=(3, 4, 9) , out_features=['stage3'] , embedding_dynamic_padding=_UpperCAmelCase , ) A_ : Optional[int] = ViTHybridConfig(backbone_config=_UpperCAmelCase , image_size=384 , num_labels=1000 ) A_ : Union[str, Any] = False # load original model from timm A_ : List[Any] = timm.create_model(_UpperCAmelCase , pretrained=_UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ : Tuple = timm_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) A_ : Any = create_rename_keys(_UpperCAmelCase , _UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : Union[str, Any] = 'huggingface/label-files' A_ : Dict = 'imagenet-1k-id2label.json' A_ : List[str] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) A_ : str = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} A_ : Any = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": A_ : List[Any] = ViTHybridModel(_UpperCAmelCase ).eval() else: A_ : str = ViTHybridForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # create image processor A_ : Dict = create_transform(**resolve_data_config({} , model=_UpperCAmelCase ) ) A_ : List[str] = transform.transforms A_ : List[str] = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } A_ : Tuple = ViTHybridImageProcessor( do_resize=_UpperCAmelCase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_UpperCAmelCase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=_UpperCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) A_ : Optional[Any] = prepare_img() A_ : Any = transform(_UpperCAmelCase ).unsqueeze(0 ) A_ : Dict = processor(_UpperCAmelCase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase ) # verify logits with torch.no_grad(): A_ : List[Any] = model(_UpperCAmelCase ) A_ : List[str] = outputs.logits print('Predicted class:' , logits.argmax(-1 ).item() ) if base_model: A_ : Union[str, Any] = timm_model.forward_features(_UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_UpperCAmelCase , outputs.pooler_output , atol=1E-3 ) else: A_ : Tuple = timm_model(_UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase , outputs.logits , atol=1E-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(f"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: print(f"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(f"""ybelkada/{vit_name}""" ) processor.push_to_hub(f"""ybelkada/{vit_name}""" ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_r50_s16_384', type=str, help='Name of the hybrid ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) lowerCamelCase_ : List[str] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
286
1
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[str] = MobileBertConfig.from_json_file(_UpperCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) A_ : List[Any] = MobileBertForPreTraining(_UpperCAmelCase ) # Load weights from tf checkpoint A_ : List[str] = load_tf_weights_in_mobilebert(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--mobilebert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained MobileBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) lowerCamelCase_ : List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
286
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" A_ : str = False if num < 0: A_ : Dict = True A_ : Union[str, Any] = -num A_ : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(_UpperCAmelCase ) for e in binary ) return "0b" + "".join(str(_UpperCAmelCase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" _validate_point(_UpperCAmelCase ) _validate_point(_UpperCAmelCase ) if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): raise ValueError('Both points must be in the same n-dimensional space' ) return float(sum(abs(a - b ) for a, b in zip(_UpperCAmelCase , _UpperCAmelCase ) ) ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if point: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): for item in point: if not isinstance(_UpperCAmelCase , (int, float) ): A_ : Optional[Any] = ( 'Expected a list of numbers as input, found ' f"""{type(_UpperCAmelCase ).__name__}""" ) raise TypeError(_UpperCAmelCase ) else: A_ : int = f"""Expected a list of numbers as input, found {type(_UpperCAmelCase ).__name__}""" raise TypeError(_UpperCAmelCase ) else: raise ValueError('Missing an input' ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" _validate_point(_UpperCAmelCase ) _validate_point(_UpperCAmelCase ) if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): raise ValueError('Both points must be in the same n-dimensional space' ) return float(sum(abs(x - y ) for x, y in zip(_UpperCAmelCase , _UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCamelCase_ : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Tuple = ['MLukeTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys lowerCamelCase_ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase = 10 , _UpperCAmelCase = 1000 , _UpperCAmelCase = True ): """simple docstring""" assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError('Invalid value for min_val or max_val (min_value < max_value)' ) return min_val if option else max_val def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return int((number_a + number_a) / 2 ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError('argument value for lower and higher must be(lower > higher)' ) if not lower < to_guess < higher: raise ValueError( 'guess value must be within the range of lower and higher value' ) def answer(_UpperCAmelCase ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print('started...' ) A_ : List[str] = lower A_ : Tuple = higher A_ : Optional[int] = [] while True: A_ : Tuple = get_avg(_UpperCAmelCase , _UpperCAmelCase ) last_numbers.append(_UpperCAmelCase ) if answer(_UpperCAmelCase ) == "low": A_ : Optional[int] = number elif answer(_UpperCAmelCase ) == "high": A_ : str = number else: break print(f"""guess the number : {last_numbers[-1]}""" ) print(f"""details : {last_numbers!s}""" ) def UpperCAmelCase__ ( ): """simple docstring""" A_ : Dict = int(input('Enter lower value : ' ).strip() ) A_ : List[Any] = int(input('Enter high value : ' ).strip() ) A_ : List[Any] = int(input('Enter value to guess : ' ).strip() ) guess_the_number(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": main()
286
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers lowerCamelCase_ : List[str] = [int(0.5 * n * (n + 1)) for n in range(1, 1_01)] def UpperCAmelCase__ ( ): """simple docstring""" A_ : Union[str, Any] = os.path.dirname(os.path.realpath(_UpperCAmelCase ) ) A_ : Tuple = os.path.join(_UpperCAmelCase , 'words.txt' ) A_ : List[Any] = '' with open(_UpperCAmelCase ) as f: A_ : int = f.readline() A_ : Optional[Any] = [word.strip('"' ) for word in words.strip('\r\n' ).split(',' )] A_ : Dict = [ word for word in [sum(ord(_UpperCAmelCase ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
286
1
"""simple docstring""" import random def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = False ): """simple docstring""" A_ : dict = {i: [] for i in range(_UpperCAmelCase )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(_UpperCAmelCase ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(_UpperCAmelCase ): for j in range(i + 1 , _UpperCAmelCase ): if random.random() < probability: graph[i].append(_UpperCAmelCase ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(_UpperCAmelCase ) return graph def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return { i: [j for j in range(_UpperCAmelCase ) if i != j] for i in range(_UpperCAmelCase ) } if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ : List[str] = {'configuration_xlnet': ['XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : str = ['XLNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[str] = ['XLNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ 'XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLNetForMultipleChoice', 'XLNetForQuestionAnswering', 'XLNetForQuestionAnsweringSimple', 'XLNetForSequenceClassification', 'XLNetForTokenClassification', 'XLNetLMHeadModel', 'XLNetModel', 'XLNetPreTrainedModel', 'load_tf_weights_in_xlnet', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Union[str, Any] = [ 'TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLNetForMultipleChoice', 'TFXLNetForQuestionAnsweringSimple', 'TFXLNetForSequenceClassification', 'TFXLNetForTokenClassification', 'TFXLNetLMHeadModel', 'TFXLNetMainLayer', 'TFXLNetModel', 'TFXLNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
1
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = (DPMSolverSinglestepScheduler,) lowercase_ : List[str] = (("""num_inference_steps""", 25),) def lowerCamelCase_ ( self , **snake_case_ ): """simple docstring""" A_ : List[Any] = { 'num_train_timesteps': 1_0_0_0, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf' ), 'variance_type': None, } config.update(**snake_case_ ) return config def lowerCamelCase_ ( self , snake_case_=0 , **snake_case_ ): """simple docstring""" A_ : List[Any] = dict(self.forward_default_kwargs ) A_ : int = kwargs.pop('num_inference_steps' , snake_case_ ) A_ : Optional[int] = self.dummy_sample A_ : Tuple = 0.1 * sample A_ : Optional[Any] = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: A_ : List[str] = self.get_scheduler_config(**snake_case_ ) A_ : Union[str, Any] = scheduler_class(**snake_case_ ) scheduler.set_timesteps(snake_case_ ) # copy over dummy past residuals A_ : Union[str, Any] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(snake_case_ ) A_ : List[str] = scheduler_class.from_pretrained(snake_case_ ) new_scheduler.set_timesteps(snake_case_ ) # copy over dummy past residuals A_ : int = dummy_past_residuals[: new_scheduler.config.solver_order] A_ , A_ : Tuple = sample, sample for t in range(snake_case_ , time_step + scheduler.config.solver_order + 1 ): A_ : Union[str, Any] = scheduler.step(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ).prev_sample A_ : str = new_scheduler.step(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCamelCase_ ( self ): """simple docstring""" pass def lowerCamelCase_ ( self , snake_case_=0 , **snake_case_ ): """simple docstring""" A_ : Optional[int] = dict(self.forward_default_kwargs ) A_ : str = kwargs.pop('num_inference_steps' , snake_case_ ) A_ : str = self.dummy_sample A_ : Optional[Any] = 0.1 * sample A_ : int = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: A_ : Dict = self.get_scheduler_config() A_ : Dict = scheduler_class(**snake_case_ ) scheduler.set_timesteps(snake_case_ ) # copy over dummy past residuals (must be after setting timesteps) A_ : int = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(snake_case_ ) A_ : Union[str, Any] = scheduler_class.from_pretrained(snake_case_ ) # copy over dummy past residuals new_scheduler.set_timesteps(snake_case_ ) # copy over dummy past residual (must be after setting timesteps) A_ : Dict = dummy_past_residuals[: new_scheduler.config.solver_order] A_ : Any = scheduler.step(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ).prev_sample A_ : Dict = new_scheduler.step(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCamelCase_ ( self , snake_case_=None , **snake_case_ ): """simple docstring""" if scheduler is None: A_ : str = self.scheduler_classes[0] A_ : Dict = self.get_scheduler_config(**snake_case_ ) A_ : int = scheduler_class(**snake_case_ ) A_ : Dict = self.scheduler_classes[0] A_ : Optional[Any] = self.get_scheduler_config(**snake_case_ ) A_ : int = scheduler_class(**snake_case_ ) A_ : Optional[Any] = 1_0 A_ : Any = self.dummy_model() A_ : Optional[int] = self.dummy_sample_deter scheduler.set_timesteps(snake_case_ ) for i, t in enumerate(scheduler.timesteps ): A_ : Union[str, Any] = model(snake_case_ , snake_case_ ) A_ : List[str] = scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample return sample def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) A_ : Any = 5_0 A_ : Union[str, Any] = self.dummy_model() A_ : Tuple = self.dummy_sample_deter scheduler.set_timesteps(snake_case_ ) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:] ): A_ : Dict = model(snake_case_ , snake_case_ ) A_ : int = scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample A_ : Any = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.25_74 ) < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : List[str] = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) A_ : str = self.full_loop(scheduler=snake_case_ ) A_ : List[str] = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.27_91 ) < 1E-3 A_ : Union[str, Any] = DEISMultistepScheduler.from_config(scheduler.config ) A_ : Union[str, Any] = DPMSolverMultistepScheduler.from_config(scheduler.config ) A_ : Optional[Any] = UniPCMultistepScheduler.from_config(scheduler.config ) A_ : Union[str, Any] = DPMSolverSinglestepScheduler.from_config(scheduler.config ) A_ : Tuple = self.full_loop(scheduler=snake_case_ ) A_ : Dict = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.27_91 ) < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" self.check_over_configs(thresholding=snake_case_ ) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=snake_case_ , prediction_type=snake_case_ , sample_max_value=snake_case_ , algorithm_type='dpmsolver++' , solver_order=snake_case_ , solver_type=snake_case_ , ) def lowerCamelCase_ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=snake_case_ , solver_type=snake_case_ , prediction_type=snake_case_ , algorithm_type=snake_case_ , ) A_ : Optional[Any] = self.full_loop( solver_order=snake_case_ , solver_type=snake_case_ , prediction_type=snake_case_ , algorithm_type=snake_case_ , ) assert not torch.isnan(snake_case_ ).any(), "Samples have nan numbers" def lowerCamelCase_ ( self ): """simple docstring""" self.check_over_configs(lower_order_final=snake_case_ ) self.check_over_configs(lower_order_final=snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" self.check_over_configs(lambda_min_clipped=-float('inf' ) ) self.check_over_configs(lambda_min_clipped=-5.1 ) def lowerCamelCase_ ( self ): """simple docstring""" self.check_over_configs(variance_type=snake_case_ ) self.check_over_configs(variance_type='learned_range' ) def lowerCamelCase_ ( self ): """simple docstring""" for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=snake_case_ , time_step=0 ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = self.full_loop() A_ : str = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.27_91 ) < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = self.full_loop(use_karras_sigmas=snake_case_ ) A_ : int = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.22_48 ) < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : str = self.full_loop(prediction_type='v_prediction' ) A_ : Optional[int] = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.14_53 ) < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.full_loop(prediction_type='v_prediction' , use_karras_sigmas=snake_case_ ) A_ : int = torch.mean(torch.abs(snake_case_ ) ) assert abs(result_mean.item() - 0.06_49 ) < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.scheduler_classes[0] A_ : Optional[int] = self.get_scheduler_config(thresholding=snake_case_ , dynamic_thresholding_ratio=0 ) A_ : int = scheduler_class(**snake_case_ ) A_ : str = 1_0 A_ : int = self.dummy_model() A_ : str = self.dummy_sample_deter.half() scheduler.set_timesteps(snake_case_ ) for i, t in enumerate(scheduler.timesteps ): A_ : Tuple = model(snake_case_ , snake_case_ ) A_ : Tuple = scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample assert sample.dtype == torch.floataa
286
"""simple docstring""" import torch from diffusers import DiffusionPipeline class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" super().__init__() self.register_modules(unet=snake_case_ , scheduler=snake_case_ ) def __call__( self ): """simple docstring""" A_ : Optional[Any] = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) A_ : List[str] = 1 A_ : List[str] = self.unet(snake_case_ , snake_case_ ).sample A_ : Optional[int] = self.scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample A_ : List[Any] = scheduler_output - scheduler_output + torch.ones_like(snake_case_ ) return result
286
1
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Union[str, Any] = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all BART models at https://huggingface.co/models?filter=bart lowerCamelCase_ : Tuple = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, } lowerCamelCase_ : List[str] = { 'facebook/bart-base': 10_24, 'facebook/bart-large': 10_24, 'facebook/bart-large-mnli': 10_24, 'facebook/bart-large-cnn': 10_24, 'facebook/bart-large-xsum': 10_24, 'yjernite/bart_eli5': 10_24, } @lru_cache() def UpperCAmelCase__ ( ): """simple docstring""" A_ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) A_ : Any = bs[:] A_ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_UpperCAmelCase ) cs.append(2**8 + n ) n += 1 A_ : Tuple = [chr(_UpperCAmelCase ) for n in cs] return dict(zip(_UpperCAmelCase , _UpperCAmelCase ) ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = set() A_ : List[str] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) A_ : Any = char return pairs class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[Any] = VOCAB_FILES_NAMES lowercase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowercase_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : Union[str, Any] = ["""input_ids""", """attention_mask"""] def __init__( self , snake_case_ , snake_case_ , snake_case_="replace" , snake_case_="<s>" , snake_case_="</s>" , snake_case_="</s>" , snake_case_="<s>" , snake_case_="<unk>" , snake_case_="<pad>" , snake_case_="<mask>" , snake_case_=False , **snake_case_ , ): """simple docstring""" A_ : List[str] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else bos_token A_ : List[str] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else eos_token A_ : List[Any] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else sep_token A_ : int = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else cls_token A_ : List[Any] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else unk_token A_ : List[Any] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it A_ : Optional[int] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token super().__init__( errors=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , cls_token=snake_case_ , pad_token=snake_case_ , mask_token=snake_case_ , add_prefix_space=snake_case_ , **snake_case_ , ) with open(snake_case_ , encoding='utf-8' ) as vocab_handle: A_ : List[Any] = json.load(snake_case_ ) A_ : Dict = {v: k for k, v in self.encoder.items()} A_ : str = errors # how to handle errors in decoding A_ : List[str] = bytes_to_unicode() A_ : Optional[int] = {v: k for k, v in self.byte_encoder.items()} with open(snake_case_ , encoding='utf-8' ) as merges_handle: A_ : Dict = merges_handle.read().split('\n' )[1:-1] A_ : List[Any] = [tuple(merge.split() ) for merge in bpe_merges] A_ : List[Any] = dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) A_ : Optional[int] = {} A_ : Union[str, Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions A_ : Union[str, Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property def lowerCamelCase_ ( self ): """simple docstring""" return len(self.encoder ) def lowerCamelCase_ ( self ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if token in self.cache: return self.cache[token] A_ : str = tuple(snake_case_ ) A_ : List[str] = get_pairs(snake_case_ ) if not pairs: return token while True: A_ : Dict = min(snake_case_ , key=lambda snake_case_ : self.bpe_ranks.get(snake_case_ , float('inf' ) ) ) if bigram not in self.bpe_ranks: break A_ , A_ : Any = bigram A_ : str = [] A_ : List[Any] = 0 while i < len(snake_case_ ): try: A_ : Union[str, Any] = word.index(snake_case_ , snake_case_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) A_ : Union[str, Any] = j if word[i] == first and i < len(snake_case_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 A_ : int = tuple(snake_case_ ) A_ : str = new_word if len(snake_case_ ) == 1: break else: A_ : Optional[Any] = get_pairs(snake_case_ ) A_ : List[Any] = ' '.join(snake_case_ ) A_ : Optional[int] = word return word def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Optional[Any] = [] for token in re.findall(self.pat , snake_case_ ): A_ : Optional[Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(snake_case_ ).split(' ' ) ) return bpe_tokens def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.encoder.get(snake_case_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.decoder.get(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[Any] = ''.join(snake_case_ ) A_ : Optional[int] = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def lowerCamelCase_ ( self , snake_case_ , snake_case_ = None ): """simple docstring""" if not os.path.isdir(snake_case_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return A_ : Union[str, Any] = os.path.join( snake_case_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) A_ : Tuple = os.path.join( snake_case_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(snake_case_ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case_ , ensure_ascii=snake_case_ ) + '\n' ) A_ : Any = 0 with open(snake_case_ , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case_ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) A_ : int = token_index writer.write(' '.join(snake_case_ ) + '\n' ) index += 1 return vocab_file, merge_file def lowerCamelCase_ ( self , snake_case_ , snake_case_ = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] A_ : str = [self.cls_token_id] A_ : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCamelCase_ ( self , snake_case_ , snake_case_ = None , snake_case_ = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) if token_ids_a is None: return [1] + ([0] * len(snake_case_ )) + [1] return [1] + ([0] * len(snake_case_ )) + [1, 1] + ([0] * len(snake_case_ )) + [1] def lowerCamelCase_ ( self , snake_case_ , snake_case_ = None ): """simple docstring""" A_ : Tuple = [self.sep_token_id] A_ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , **snake_case_ ): """simple docstring""" A_ : int = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(snake_case_ ) > 0 and not text[0].isspace()): A_ : Optional[int] = ' ' + text return (text, kwargs)
286
"""simple docstring""" from heapq import heappop, heappush import numpy as np def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" A_ , A_ : List[str] = grid.shape A_ : Optional[int] = [-1, 1, 0, 0] A_ : str = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] A_ , A_ : List[Any] = [(0, source)], set() A_ : Optional[Any] = np.full((rows, cols) , np.inf ) A_ : int = 0 A_ : Optional[int] = np.empty((rows, cols) , dtype=_UpperCAmelCase ) A_ : Optional[int] = None while queue: ((A_) , (A_)) : str = heappop(_UpperCAmelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: A_ : int = [] while (x, y) != source: path.append((x, y) ) A_ , A_ : List[Any] = predecessors[x, y] path.append(_UpperCAmelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(_UpperCAmelCase ) ): A_ , A_ : Tuple = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: A_ : Union[str, Any] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(_UpperCAmelCase , (dist + 1, (nx, ny)) ) A_ : Optional[Any] = dist + 1 A_ : Optional[Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.25.0')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
286
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase_ : Optional[Any] = { 'huggingface/informer-tourism-monthly': ( 'https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json' ), # See all Informer models at https://huggingface.co/models?filter=informer } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """informer""" lowercase_ : str = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self , snake_case_ = None , snake_case_ = None , snake_case_ = "student_t" , snake_case_ = "nll" , snake_case_ = 1 , snake_case_ = None , snake_case_ = "mean" , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = 6_4 , snake_case_ = 3_2 , snake_case_ = 3_2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = True , snake_case_ = "gelu" , snake_case_ = 0.05 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 1_0_0 , snake_case_ = 0.02 , snake_case_=True , snake_case_ = "prob" , snake_case_ = 5 , snake_case_ = True , **snake_case_ , ): """simple docstring""" A_ : str = prediction_length A_ : List[Any] = context_length or prediction_length A_ : str = distribution_output A_ : Dict = loss A_ : Any = input_size A_ : Union[str, Any] = num_time_features A_ : Optional[Any] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] A_ : List[Any] = scaling A_ : Tuple = num_dynamic_real_features A_ : Any = num_static_real_features A_ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) A_ : Optional[int] = cardinality else: A_ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) A_ : Any = embedding_dimension else: A_ : Optional[Any] = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] A_ : int = num_parallel_samples # Transformer architecture configuration A_ : str = input_size * len(self.lags_sequence ) + self._number_of_features A_ : List[Any] = d_model A_ : Dict = encoder_attention_heads A_ : Dict = decoder_attention_heads A_ : List[Any] = encoder_ffn_dim A_ : Union[str, Any] = decoder_ffn_dim A_ : int = encoder_layers A_ : Any = decoder_layers A_ : List[Any] = dropout A_ : str = attention_dropout A_ : Tuple = activation_dropout A_ : List[str] = encoder_layerdrop A_ : List[str] = decoder_layerdrop A_ : str = activation_function A_ : Optional[int] = init_std A_ : List[Any] = use_cache # Informer A_ : Tuple = attention_type A_ : List[Any] = sampling_factor A_ : Optional[int] = distil super().__init__(is_encoder_decoder=snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
286
1
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
"""simple docstring""" import os def UpperCAmelCase__ ( ): """simple docstring""" A_ : Any = os.path.join(os.path.dirname(_UpperCAmelCase ) , 'num.txt' ) with open(_UpperCAmelCase ) as file_hand: return str(sum(int(_UpperCAmelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
286
1
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowercase_ : Union[str, Any] = IFInpaintingSuperResolutionPipeline lowercase_ : Optional[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} lowercase_ : int = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) lowercase_ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def lowerCamelCase_ ( self ): """simple docstring""" return self._get_superresolution_dummy_components() def lowerCamelCase_ ( self , snake_case_ , snake_case_=0 ): """simple docstring""" if str(snake_case_ ).startswith('mps' ): A_ : Tuple = torch.manual_seed(snake_case_ ) else: A_ : Any = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) A_ : Any = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ : Optional[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ : List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def lowerCamelCase_ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def lowerCamelCase_ ( self ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def lowerCamelCase_ ( self ): """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def lowerCamelCase_ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def lowerCamelCase_ ( self ): """simple docstring""" self._test_save_load_local() def lowerCamelCase_ ( self ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
286
"""simple docstring""" import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger lowerCamelCase_ : Dict = get_logger(__name__) lowerCamelCase_ : List[str] = r'\n Args:\n input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam\n search or log softmax for each vocabulary token when using beam search\n kwargs (`Dict[str, Any]`, *optional*):\n Additional logits processor specific kwargs.\n\n Return:\n `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores.\n\n' class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ): """simple docstring""" for processor in self: A_ : Tuple = inspect.signature(processor.__call__ ).parameters if len(snake_case_ ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( F"""Make sure that all the required parameters: {list(function_args.keys() )} for """ F"""{processor.__class__} are passed to the logits processor.""" ) A_ : Tuple = processor(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) else: A_ : Optional[Any] = processor(snake_case_ , snake_case_ , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or not (temperature > 0): raise ValueError(F"""`temperature` has to be a strictly positive float, but is {temperature}""" ) A_ : Optional[int] = temperature def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = scores / self.temperature return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or (top_p < 0 or top_p > 1.0): raise ValueError(F"""`top_p` has to be a float > 0 and < 1, but is {top_p}""" ) if not isinstance(snake_case_ , snake_case_ ) or (min_tokens_to_keep < 1): raise ValueError(F"""`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}""" ) A_ : str = top_p A_ : Union[str, Any] = filter_value A_ : int = min_tokens_to_keep def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : Tuple = lax.top_k(snake_case_ , scores.shape[-1] ) A_ : List[Any] = jnp.full_like(snake_case_ , self.filter_value ) A_ : List[str] = jax.nn.softmax(snake_case_ , axis=-1 ).cumsum(axis=-1 ) A_ : Optional[int] = cumulative_probs < self.top_p # include the token that is higher than top_p as well A_ : Union[str, Any] = jnp.roll(snake_case_ , 1 ) score_mask |= score_mask.at[:, 0].set(snake_case_ ) # min tokens to keep A_ : int = score_mask.at[:, : self.min_tokens_to_keep].set(snake_case_ ) A_ : Optional[Any] = jnp.where(snake_case_ , snake_case_ , snake_case_ ) A_ : List[Any] = jax.lax.sort_key_val(snake_case_ , snake_case_ )[-1] return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or top_k <= 0: raise ValueError(F"""`top_k` has to be a strictly positive integer, but is {top_k}""" ) A_ : str = max(snake_case_ , snake_case_ ) A_ : Union[str, Any] = filter_value def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : int = scores.shape A_ : Tuple = jnp.full(batch_size * vocab_size , self.filter_value ) A_ : Union[str, Any] = min(self.top_k , scores.shape[-1] ) # Safety check A_ , A_ : Dict = lax.top_k(snake_case_ , snake_case_ ) A_ : Optional[int] = jnp.broadcast_to((jnp.arange(snake_case_ ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() A_ : int = topk_scores.flatten() A_ : Any = topk_indices.flatten() + shift A_ : List[str] = next_scores_flat.at[topk_indices_flat].set(snake_case_ ) A_ : Union[str, Any] = next_scores_flat.reshape(snake_case_ , snake_case_ ) return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = bos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Union[str, Any] = 1 - jnp.bool_(cur_len - 1 ) A_ : str = jnp.where(snake_case_ , new_scores.at[:, self.bos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = max_length A_ : Optional[int] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Dict = 1 - jnp.bool_(cur_len - self.max_length + 1 ) A_ : Dict = jnp.where(snake_case_ , new_scores.at[:, self.eos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or min_length < 0: raise ValueError(F"""`min_length` has to be a positive integer, but is {min_length}""" ) if not isinstance(snake_case_ , snake_case_ ) or eos_token_id < 0: raise ValueError(F"""`eos_token_id` has to be a positive integer, but is {eos_token_id}""" ) A_ : Any = min_length A_ : List[Any] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) A_ : Optional[Any] = jnp.where(snake_case_ , scores.at[:, self.eos_token_id].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) A_ : Tuple = begin_index def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = 1 - jnp.bool_(cur_len - self.begin_index ) A_ : int = jnp.where(snake_case_ , scores.at[:, self.begin_suppress_tokens].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = scores.at[..., self.suppress_tokens].set(-float('inf' ) ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Any = dict(snake_case_ ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. A_ : Tuple = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: A_ : Tuple = force_token_array.at[index].set(snake_case_ ) A_ : Any = jnp.intaa(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" def _force_token(snake_case_ ): A_ : List[Any] = scores.shape[0] A_ : Any = self.force_token_array[generation_idx] A_ : Tuple = jnp.ones_like(snake_case_ , dtype=scores.dtype ) * -float('inf' ) A_ : List[Any] = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) A_ : int = lax.dynamic_update_slice(snake_case_ , snake_case_ , (0, current_token) ) return new_scores A_ : int = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(snake_case_ ) , lambda: scores , ) , ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = generate_config.eos_token_id A_ : Optional[int] = generate_config.no_timestamps_token_id A_ : List[str] = generate_config.no_timestamps_token_id + 1 A_ : Any = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(snake_case_ , 'max_initial_timestamp_index' ): A_ : List[Any] = generate_config.max_initial_timestamp_index else: A_ : Any = model_config.vocab_size if self.max_initial_timestamp_index is None: A_ : Optional[Any] = model_config.vocab_size def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[str] = scores.at[:, self.no_timestamps_token_id].set(-float('inf' ) ) def handle_pairs(snake_case_ , snake_case_ ): A_ : Any = jnp.where((cur_len - self.begin_index) >= 1 , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , snake_case_ , ) A_ : Tuple = jnp.where((cur_len - self.begin_index) < 2 , snake_case_ , snake_case_ ) A_ : Any = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , snake_case_ , snake_case_ , ) return jnp.where( snake_case_ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float('inf' ) ) , scores_k.at[: self.eos_token_id].set(-float('inf' ) ) , ) , snake_case_ , ) A_ : Tuple = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) A_ : Optional[Any] = jnp.where(cur_len == self.begin_index , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , snake_case_ , ) A_ : int = self.timestamp_begin + self.max_initial_timestamp_index A_ : List[Any] = jnp.where( snake_case_ , scores.at[:, last_allowed + 1 :].set(-float('inf' ) ) , snake_case_ , ) # if sum of probability over timestamps is above any other token, sample timestamp A_ : Any = jax.nn.log_softmax(snake_case_ , axis=-1 ) def handle_cumulative_probs(snake_case_ , snake_case_ ): A_ : Dict = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) A_ : Optional[Any] = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float('inf' ) ) , snake_case_ , ) A_ : Union[str, Any] = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) return scores
286
1
"""simple docstring""" from __future__ import annotations from collections.abc import Generator def UpperCAmelCase__ ( ): """simple docstring""" A_ : dict[int, int] = {} A_ : Optional[Any] = 2 while True: A_ : Union[str, Any] = factor_map.pop(_UpperCAmelCase , _UpperCAmelCase ) if factor: A_ : Optional[Any] = factor + prime while x in factor_map: x += factor A_ : Dict = factor else: A_ : List[str] = prime yield prime prime += 1 def UpperCAmelCase__ ( _UpperCAmelCase = 1E10 ): """simple docstring""" A_ : Optional[int] = sieve() A_ : List[str] = 1 while True: A_ : List[Any] = next(_UpperCAmelCase ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(_UpperCAmelCase ) n += 2 if __name__ == "__main__": print(solution())
286
"""simple docstring""" import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = R'\w+[.]\d+' A_ : int = re.findall(_UpperCAmelCase , _UpperCAmelCase ) for pat in pats: A_ : Optional[int] = key.replace(_UpperCAmelCase , '_'.join(pat.split('.' ) ) ) return key def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): A_ : Union[str, Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: A_ : List[str] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: A_ : Optional[Any] = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer A_ : int = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: A_ : str = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer A_ : Optional[Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": A_ : Optional[Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight A_ : Tuple = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias A_ : Optional[int] = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=42 ): """simple docstring""" A_ : int = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params A_ : Union[str, Any] = flax_model.init_weights(PRNGKey(_UpperCAmelCase ) ) A_ : Optional[Any] = flatten_dict(_UpperCAmelCase ) A_ : Tuple = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): A_ : Any = rename_key(_UpperCAmelCase ) A_ : List[str] = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters A_ , A_ : Union[str, Any] = rename_key_and_reshape_tensor(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown A_ : str = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase )
286
1
"""simple docstring""" import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ , snake_case_=1_3 , snake_case_=3_0 , snake_case_=2 , snake_case_=3 , snake_case_=True , snake_case_=True , snake_case_=3_2 , snake_case_=5 , snake_case_=4 , snake_case_=3_7 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=1_0 , snake_case_=0.02 , snake_case_=3 , snake_case_=0.6 , snake_case_=None , ): """simple docstring""" A_ : Optional[Any] = parent A_ : Union[str, Any] = batch_size A_ : str = image_size A_ : int = patch_size A_ : Any = num_channels A_ : List[Any] = is_training A_ : Tuple = use_labels A_ : List[Any] = hidden_size A_ : Optional[Any] = num_hidden_layers A_ : int = num_attention_heads A_ : Optional[int] = intermediate_size A_ : int = hidden_act A_ : Tuple = hidden_dropout_prob A_ : Optional[Any] = attention_probs_dropout_prob A_ : Tuple = type_sequence_label_size A_ : int = initializer_range A_ : str = mask_ratio A_ : int = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) A_ : str = (image_size // patch_size) ** 2 A_ : Any = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Union[str, Any] = None if self.use_labels: A_ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Any = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self ): """simple docstring""" return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[int] = ViTMAEModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : int = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Any = ViTMAEForPreTraining(snake_case_ ) model.to(snake_case_ ) model.eval() A_ : int = model(snake_case_ ) A_ : List[Any] = (self.image_size // self.patch_size) ** 2 A_ : Optional[int] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images A_ : Tuple = 1 A_ : Any = ViTMAEForPreTraining(snake_case_ ) model.to(snake_case_ ) model.eval() A_ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(snake_case_ ) A_ : Optional[Any] = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = self.prepare_config_and_inputs() A_ , A_ , A_ : int = config_and_inputs A_ : int = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowercase_ : Tuple = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () lowercase_ : Tuple = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} lowercase_ : List[str] = False lowercase_ : Optional[int] = False lowercase_ : Union[str, Any] = False lowercase_ : Optional[Any] = False def lowerCamelCase_ ( self ): """simple docstring""" A_ : str = ViTMAEModelTester(self ) A_ : str = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=3_7 ) def lowerCamelCase_ ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def lowerCamelCase_ ( self ): """simple docstring""" pass def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(snake_case_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_ , nn.Linear ) ) def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Dict = model_class(snake_case_ ) A_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Any = [*signature.parameters.keys()] A_ : List[str] = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" np.random.seed(2 ) A_ : Dict = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) A_ : Any = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) A_ : Tuple = torch.from_numpy(snake_case_ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument A_ : Tuple = pt_noise super().check_pt_tf_models(snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Dict = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): A_ : Optional[Any] = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) A_ : Optional[Any] = outputs[0].cpu().numpy() A_ : List[Any] = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ ) A_ : Any = model_class.from_pretrained(snake_case_ ) model.to(snake_case_ ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): A_ : List[str] = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) # Make sure we don't have nans A_ : Union[str, Any] = after_outputs[0].cpu().numpy() A_ : Union[str, Any] = 0 A_ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(snake_case_ , 1E-5 ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def lowerCamelCase_ ( self ): """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def lowerCamelCase_ ( self ): """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def lowerCamelCase_ ( self ): """simple docstring""" pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def lowerCamelCase_ ( self ): """simple docstring""" pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def lowerCamelCase_ ( self ): """simple docstring""" pass @slow def lowerCamelCase_ ( self ): """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = ViTMAEModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def UpperCAmelCase__ ( ): """simple docstring""" A_ : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCamelCase_ ( self ): """simple docstring""" return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def lowerCamelCase_ ( self ): """simple docstring""" np.random.seed(2 ) A_ : List[str] = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ).to(snake_case_ ) A_ : Tuple = self.default_image_processor A_ : Optional[Any] = prepare_img() A_ : List[str] = image_processor(images=snake_case_ , return_tensors='pt' ).to(snake_case_ ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) A_ : List[str] = ViTMAEConfig() A_ : Optional[Any] = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) A_ : List[Any] = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): A_ : int = model(**snake_case_ , noise=torch.from_numpy(snake_case_ ).to(device=snake_case_ ) ) # verify the logits A_ : Optional[Any] = torch.Size((1, 1_9_6, 7_6_8) ) self.assertEqual(outputs.logits.shape , snake_case_ ) A_ : int = torch.tensor( [[-0.05_48, -1.70_23, -0.93_25], [0.37_21, -0.56_70, -0.22_33], [0.82_35, -1.38_78, -0.35_24]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(snake_case_ ) , atol=1E-4 ) )
286
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = CustomTokenizer pass
286
1
"""simple docstring""" import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=1 ): """simple docstring""" if n_shave_prefix_segments >= 0: return ".".join(path.split('.' )[n_shave_prefix_segments:] ) else: return ".".join(path.split('.' )[:n_shave_prefix_segments] ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=0 ): """simple docstring""" A_ : List[str] = [] for old_item in old_list: A_ : Any = old_item.replace('in_layers.0' , 'norm1' ) A_ : List[Any] = new_item.replace('in_layers.2' , 'conv1' ) A_ : Union[str, Any] = new_item.replace('out_layers.0' , 'norm2' ) A_ : Dict = new_item.replace('out_layers.3' , 'conv2' ) A_ : Dict = new_item.replace('emb_layers.1' , 'time_emb_proj' ) A_ : Union[str, Any] = new_item.replace('skip_connection' , 'conv_shortcut' ) A_ : Optional[int] = shave_segments(_UpperCAmelCase , n_shave_prefix_segments=_UpperCAmelCase ) mapping.append({'old': old_item, 'new': new_item} ) return mapping def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=0 ): """simple docstring""" A_ : Optional[Any] = [] for old_item in old_list: A_ : List[Any] = old_item A_ : Any = new_item.replace('norm.weight' , 'group_norm.weight' ) A_ : int = new_item.replace('norm.bias' , 'group_norm.bias' ) A_ : Union[str, Any] = new_item.replace('proj_out.weight' , 'proj_attn.weight' ) A_ : List[Any] = new_item.replace('proj_out.bias' , 'proj_attn.bias' ) A_ : List[str] = shave_segments(_UpperCAmelCase , n_shave_prefix_segments=_UpperCAmelCase ) mapping.append({'old': old_item, 'new': new_item} ) return mapping def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None ): """simple docstring""" assert isinstance(_UpperCAmelCase , _UpperCAmelCase ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): A_ : int = old_checkpoint[path] A_ : List[str] = old_tensor.shape[0] // 3 A_ : List[Any] = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) A_ : int = old_tensor.shape[0] // config['num_head_channels'] // 3 A_ : Optional[Any] = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) A_ , A_ , A_ : Any = old_tensor.split(channels // num_heads , dim=1 ) A_ : str = query.reshape(_UpperCAmelCase ) A_ : Union[str, Any] = key.reshape(_UpperCAmelCase ) A_ : int = value.reshape(_UpperCAmelCase ) for path in paths: A_ : Optional[int] = path['new'] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here A_ : Optional[Any] = new_path.replace('middle_block.0' , 'mid_block.resnets.0' ) A_ : Any = new_path.replace('middle_block.1' , 'mid_block.attentions.0' ) A_ : Dict = new_path.replace('middle_block.2' , 'mid_block.resnets.1' ) if additional_replacements is not None: for replacement in additional_replacements: A_ : Optional[int] = new_path.replace(replacement['old'] , replacement['new'] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: A_ : List[Any] = old_checkpoint[path['old']][:, :, 0] else: A_ : Any = old_checkpoint[path['old']] def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = {} A_ : Tuple = checkpoint['time_embed.0.weight'] A_ : List[Any] = checkpoint['time_embed.0.bias'] A_ : List[str] = checkpoint['time_embed.2.weight'] A_ : List[str] = checkpoint['time_embed.2.bias'] A_ : Dict = checkpoint['input_blocks.0.0.weight'] A_ : str = checkpoint['input_blocks.0.0.bias'] A_ : Tuple = checkpoint['out.0.weight'] A_ : int = checkpoint['out.0.bias'] A_ : Any = checkpoint['out.2.weight'] A_ : str = checkpoint['out.2.bias'] # Retrieves the keys for the input blocks only A_ : int = len({'.'.join(layer.split('.' )[:2] ) for layer in checkpoint if 'input_blocks' in layer} ) A_ : str = { layer_id: [key for key in checkpoint if f"""input_blocks.{layer_id}""" in key] for layer_id in range(_UpperCAmelCase ) } # Retrieves the keys for the middle blocks only A_ : Optional[int] = len({'.'.join(layer.split('.' )[:2] ) for layer in checkpoint if 'middle_block' in layer} ) A_ : Optional[Any] = { layer_id: [key for key in checkpoint if f"""middle_block.{layer_id}""" in key] for layer_id in range(_UpperCAmelCase ) } # Retrieves the keys for the output blocks only A_ : List[str] = len({'.'.join(layer.split('.' )[:2] ) for layer in checkpoint if 'output_blocks' in layer} ) A_ : Optional[int] = { layer_id: [key for key in checkpoint if f"""output_blocks.{layer_id}""" in key] for layer_id in range(_UpperCAmelCase ) } for i in range(1 , _UpperCAmelCase ): A_ : Optional[Any] = (i - 1) // (config['num_res_blocks'] + 1) A_ : List[str] = (i - 1) % (config['num_res_blocks'] + 1) A_ : int = [key for key in input_blocks[i] if f"""input_blocks.{i}.0""" in key] A_ : str = [key for key in input_blocks[i] if f"""input_blocks.{i}.1""" in key] if f"""input_blocks.{i}.0.op.weight""" in checkpoint: A_ : Dict = checkpoint[ f"""input_blocks.{i}.0.op.weight""" ] A_ : Union[str, Any] = checkpoint[ f"""input_blocks.{i}.0.op.bias""" ] continue A_ : List[str] = renew_resnet_paths(_UpperCAmelCase ) A_ : Union[str, Any] = {'old': f"""input_blocks.{i}.0""", 'new': f"""down_blocks.{block_id}.resnets.{layer_in_block_id}"""} A_ : List[str] = {'old': 'resnets.2.op', 'new': 'downsamplers.0.op'} assign_to_checkpoint( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path, resnet_op] , config=_UpperCAmelCase ) if len(_UpperCAmelCase ): A_ : Tuple = renew_attention_paths(_UpperCAmelCase ) A_ : List[Any] = { 'old': f"""input_blocks.{i}.1""", 'new': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}""", } A_ : Union[str, Any] = { f"""input_blocks.{i}.1.qkv.bias""": { 'key': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", 'query': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", 'value': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, f"""input_blocks.{i}.1.qkv.weight""": { 'key': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", 'query': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", 'value': f"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , attention_paths_to_split=_UpperCAmelCase , config=_UpperCAmelCase , ) A_ : List[Any] = middle_blocks[0] A_ : Union[str, Any] = middle_blocks[1] A_ : Union[str, Any] = middle_blocks[2] A_ : str = renew_resnet_paths(_UpperCAmelCase ) assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , config=_UpperCAmelCase ) A_ : Union[str, Any] = renew_resnet_paths(_UpperCAmelCase ) assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , config=_UpperCAmelCase ) A_ : Optional[Any] = renew_attention_paths(_UpperCAmelCase ) A_ : List[Any] = { 'middle_block.1.qkv.bias': { 'key': 'mid_block.attentions.0.key.bias', 'query': 'mid_block.attentions.0.query.bias', 'value': 'mid_block.attentions.0.value.bias', }, 'middle_block.1.qkv.weight': { 'key': 'mid_block.attentions.0.key.weight', 'query': 'mid_block.attentions.0.query.weight', 'value': 'mid_block.attentions.0.value.weight', }, } assign_to_checkpoint( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , attention_paths_to_split=_UpperCAmelCase , config=_UpperCAmelCase ) for i in range(_UpperCAmelCase ): A_ : Tuple = i // (config['num_res_blocks'] + 1) A_ : Tuple = i % (config['num_res_blocks'] + 1) A_ : Optional[Any] = [shave_segments(_UpperCAmelCase , 2 ) for name in output_blocks[i]] A_ : Union[str, Any] = {} for layer in output_block_layers: A_ , A_ : List[Any] = layer.split('.' )[0], shave_segments(_UpperCAmelCase , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(_UpperCAmelCase ) else: A_ : Any = [layer_name] if len(_UpperCAmelCase ) > 1: A_ : List[str] = [key for key in output_blocks[i] if f"""output_blocks.{i}.0""" in key] A_ : List[str] = [key for key in output_blocks[i] if f"""output_blocks.{i}.1""" in key] A_ : Optional[int] = renew_resnet_paths(_UpperCAmelCase ) A_ : List[Any] = renew_resnet_paths(_UpperCAmelCase ) A_ : Optional[Any] = {'old': f"""output_blocks.{i}.0""", 'new': f"""up_blocks.{block_id}.resnets.{layer_in_block_id}"""} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) if ["conv.weight", "conv.bias"] in output_block_list.values(): A_ : List[Any] = list(output_block_list.values() ).index(['conv.weight', 'conv.bias'] ) A_ : int = checkpoint[ f"""output_blocks.{i}.{index}.conv.weight""" ] A_ : Optional[int] = checkpoint[ f"""output_blocks.{i}.{index}.conv.bias""" ] # Clear attentions as they have been attributed above. if len(_UpperCAmelCase ) == 2: A_ : int = [] if len(_UpperCAmelCase ): A_ : Union[str, Any] = renew_attention_paths(_UpperCAmelCase ) A_ : Dict = { 'old': f"""output_blocks.{i}.1""", 'new': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}""", } A_ : Optional[Any] = { f"""output_blocks.{i}.1.qkv.bias""": { 'key': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", 'query': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", 'value': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, f"""output_blocks.{i}.1.qkv.weight""": { 'key': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", 'query': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", 'value': f"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any('qkv' in key for key in attentions ) else None , config=_UpperCAmelCase , ) else: A_ : Optional[Any] = renew_resnet_paths(_UpperCAmelCase , n_shave_prefix_segments=1 ) for path in resnet_0_paths: A_ : List[str] = '.'.join(['output_blocks', str(_UpperCAmelCase ), path['old']] ) A_ : Optional[Any] = '.'.join(['up_blocks', str(_UpperCAmelCase ), 'resnets', str(_UpperCAmelCase ), path['new']] ) A_ : Optional[int] = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": lowerCamelCase_ : Optional[int] = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the architecture.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') lowerCamelCase_ : Any = parser.parse_args() lowerCamelCase_ : Dict = torch.load(args.checkpoint_path) with open(args.config_file) as f: lowerCamelCase_ : Dict = json.loads(f.read()) lowerCamelCase_ : Tuple = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] lowerCamelCase_ : List[Any] = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: lowerCamelCase_ : Optional[Any] = DDPMScheduler.from_config('/'.join(args.checkpoint_path.split('/')[:-1])) lowerCamelCase_ : str = VQModel.from_pretrained('/'.join(args.checkpoint_path.split('/')[:-1])) lowerCamelCase_ : Optional[int] = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
286
"""simple docstring""" from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch lowerCamelCase_ : str = logging.get_logger(__name__) @add_end_docstrings( UpperCAmelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.framework == "tf": A_ : str = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": A_ : List[str] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ) else: raise ValueError('Unsupported framework' ) return masked_index def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = self.get_masked_index(snake_case_ ) A_ : str = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['input_ids'][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_=None , **snake_case_ ): """simple docstring""" if return_tensors is None: A_ : Any = self.framework A_ : Dict = self.tokenizer(snake_case_ , return_tensors=snake_case_ ) self.ensure_exactly_one_mask_token(snake_case_ ) return model_inputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Dict = self.model(**snake_case_ ) A_ : Optional[int] = model_inputs['input_ids'] return model_outputs def lowerCamelCase_ ( self , snake_case_ , snake_case_=5 , snake_case_=None ): """simple docstring""" if target_ids is not None and target_ids.shape[0] < top_k: A_ : str = target_ids.shape[0] A_ : Optional[Any] = model_outputs['input_ids'][0] A_ : List[Any] = model_outputs['logits'] if self.framework == "tf": A_ : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] A_ : Union[str, Any] = outputs.numpy() A_ : Optional[int] = outputs[0, masked_index, :] A_ : Optional[Any] = stable_softmax(snake_case_ , axis=-1 ) if target_ids is not None: A_ : Union[str, Any] = tf.gather_nd(tf.squeeze(snake_case_ , 0 ) , target_ids.reshape(-1 , 1 ) ) A_ : Optional[int] = tf.expand_dims(snake_case_ , 0 ) A_ : Any = tf.math.top_k(snake_case_ , k=snake_case_ ) A_ , A_ : str = topk.values.numpy(), topk.indices.numpy() else: A_ : int = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample A_ : Tuple = outputs[0, masked_index, :] A_ : List[str] = logits.softmax(dim=-1 ) if target_ids is not None: A_ : str = probs[..., target_ids] A_ , A_ : List[str] = probs.topk(snake_case_ ) A_ : List[Any] = [] A_ : int = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): A_ : str = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place A_ : Union[str, Any] = input_ids.numpy().copy() if target_ids is not None: A_ : str = target_ids[p].tolist() A_ : Union[str, Any] = p # Filter padding out: A_ : Any = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back A_ : Any = self.tokenizer.decode(snake_case_ , skip_special_tokens=snake_case_ ) A_ : Any = {'score': v, 'token': p, 'token_str': self.tokenizer.decode([p] ), 'sequence': sequence} row.append(snake_case_ ) result.append(snake_case_ ) if single_mask: return result[0] return result def lowerCamelCase_ ( self , snake_case_ , snake_case_=None ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : List[str] = [targets] try: A_ : Optional[int] = self.tokenizer.get_vocab() except Exception: A_ : int = {} A_ : Tuple = [] for target in targets: A_ : int = vocab.get(snake_case_ , snake_case_ ) if id_ is None: A_ : Tuple = self.tokenizer( snake_case_ , add_special_tokens=snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , max_length=1 , truncation=snake_case_ , )['input_ids'] if len(snake_case_ ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ 'We cannot replace it with anything meaningful, ignoring it' ) continue A_ : str = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) A_ : Tuple = list(set(snake_case_ ) ) if len(snake_case_ ) == 0: raise ValueError('At least one target must be provided when passed.' ) A_ : Optional[Any] = np.array(snake_case_ ) return target_ids def lowerCamelCase_ ( self , snake_case_=None , snake_case_=None ): """simple docstring""" A_ : List[str] = {} if targets is not None: A_ : Any = self.get_target_ids(snake_case_ , snake_case_ ) A_ : Optional[Any] = target_ids if top_k is not None: A_ : int = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , 'The tokenizer does not define a `mask_token`.' ) return {}, {}, postprocess_params def __call__( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : List[str] = super().__call__(snake_case_ , **snake_case_ ) if isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1: return outputs[0] return outputs
286
1
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": lowerCamelCase_ : Tuple = pd.read_csv('sample_data.csv', header=None) lowerCamelCase_ : int = df.shape[:1][0] # If you're using some other dataset input the target column lowerCamelCase_ : Union[str, Any] = df.iloc[:, 1:2] lowerCamelCase_ : Dict = actual_data.values.reshape(len_data, 1) lowerCamelCase_ : int = MinMaxScaler().fit_transform(actual_data) lowerCamelCase_ : Dict = 10 lowerCamelCase_ : Tuple = 5 lowerCamelCase_ : Any = 20 lowerCamelCase_ : Dict = len_data - periods * look_back lowerCamelCase_ : Optional[int] = actual_data[:division] lowerCamelCase_ : Union[str, Any] = actual_data[division - look_back :] lowerCamelCase_ , lowerCamelCase_ : Dict = [], [] lowerCamelCase_ , lowerCamelCase_ : Union[str, Any] = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) lowerCamelCase_ : Optional[Any] = np.array(train_x) lowerCamelCase_ : Union[str, Any] = np.array(test_x) lowerCamelCase_ : Dict = np.array([list(i.ravel()) for i in train_y]) lowerCamelCase_ : List[Any] = np.array([list(i.ravel()) for i in test_y]) lowerCamelCase_ : List[str] = Sequential() model.add(LSTM(1_28, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(1_28, 1))) model.add(Dense(forward_days)) model.compile(loss='mean_squared_error', optimizer='adam') lowerCamelCase_ : Dict = model.fit( x_train, y_train, epochs=1_50, verbose=1, shuffle=True, batch_size=4 ) lowerCamelCase_ : Any = model.predict(x_test)
286
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return round(float(moles / volume ) * nfactor ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return round(float((moles * 0.0_821 * temperature) / (volume) ) ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return round(float((moles * 0.0_821 * temperature) / (pressure) ) ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" return round(float((pressure * volume) / (0.0_821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) A_ : Any = hex_num[0] == '-' if is_negative: A_ : Optional[Any] = hex_num[1:] try: A_ : Tuple = int(_UpperCAmelCase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) A_ : Union[str, Any] = '' while int_num > 0: A_ : Optional[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase_ : str = { 'configuration_table_transformer': [ 'TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TableTransformerConfig', 'TableTransformerOnnxConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Optional[int] = [ 'TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TableTransformerForObjectDetection', 'TableTransformerModel', 'TableTransformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase_ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ , A_ : Dict = len(_UpperCAmelCase ), len(grid[0] ) if ( min(_UpperCAmelCase , _UpperCAmelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) A_ : Optional[int] = 0 count += depth_first_search(_UpperCAmelCase , row + 1 , _UpperCAmelCase , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , row - 1 , _UpperCAmelCase , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , _UpperCAmelCase , col + 1 , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , _UpperCAmelCase , col - 1 , _UpperCAmelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Any = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """xmod""" def __init__( self , snake_case_=3_0_5_2_2 , snake_case_=7_6_8 , snake_case_=1_2 , snake_case_=1_2 , snake_case_=3_0_7_2 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , snake_case_=False , snake_case_=2 , snake_case_=False , snake_case_=True , snake_case_=True , snake_case_=("en_XX",) , snake_case_=None , **snake_case_ , ): """simple docstring""" super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) A_ : Union[str, Any] = vocab_size A_ : Any = hidden_size A_ : List[str] = num_hidden_layers A_ : Tuple = num_attention_heads A_ : int = hidden_act A_ : Any = intermediate_size A_ : Any = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Union[str, Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : List[str] = initializer_range A_ : Any = layer_norm_eps A_ : Optional[Any] = position_embedding_type A_ : int = use_cache A_ : Dict = classifier_dropout A_ : int = pre_norm A_ : Optional[Any] = adapter_reduction_factor A_ : List[Any] = adapter_layer_norm A_ : int = adapter_reuse_layer_norm A_ : Dict = ln_before_adapter A_ : List[str] = list(snake_case_ ) A_ : Union[str, Any] = default_language class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @property def lowerCamelCase_ ( self ): """simple docstring""" if self.task == "multiple-choice": A_ : Dict = {0: 'batch', 1: 'choice', 2: 'sequence'} else: A_ : int = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) A_ : Any = hex_num[0] == '-' if is_negative: A_ : Optional[Any] = hex_num[1:] try: A_ : Tuple = int(_UpperCAmelCase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) A_ : Union[str, Any] = '' while int_num > 0: A_ : Optional[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
1
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
"""simple docstring""" from copy import deepcopy class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None , snake_case_ = None ): """simple docstring""" if arr is None and size is not None: A_ : Union[str, Any] = size A_ : List[str] = [0] * size elif arr is not None: self.init(snake_case_ ) else: raise ValueError('Either arr or size must be specified' ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = len(snake_case_ ) A_ : Optional[int] = deepcopy(snake_case_ ) for i in range(1 , self.size ): A_ : Optional[Any] = self.next_(snake_case_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): A_ : Optional[int] = self.next_(snake_case_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index + (index & (-index)) @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index - (index & (-index)) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value A_ : List[str] = self.next_(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" self.add(snake_case_ , value - self.get(snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if right == 0: return 0 A_ : Any = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] A_ : Tuple = self.prev(snake_case_ ) return result def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" return self.prefix(snake_case_ ) - self.prefix(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.query(snake_case_ , index + 1 ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" value -= self.tree[0] if value < 0: return -1 A_ : List[Any] = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 A_ : Tuple = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowercase_ : Tuple = StableDiffusionInstructPixaPixPipeline lowercase_ : Any = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} lowercase_ : Dict = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowercase_ : int = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase_ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase_ ( self ): """simple docstring""" torch.manual_seed(0 ) A_ : Dict = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) A_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=snake_case_ ) torch.manual_seed(0 ) A_ : Tuple = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) A_ : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) A_ : int = CLIPTextModel(snake_case_ ) A_ : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) A_ : Tuple = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowerCamelCase_ ( self , snake_case_ , snake_case_=0 ): """simple docstring""" A_ : Any = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] A_ : Optional[Any] = Image.fromarray(np.uinta(snake_case_ ) ).convert('RGB' ) if str(snake_case_ ).startswith('mps' ): A_ : Optional[Any] = torch.manual_seed(snake_case_ ) else: A_ : int = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) A_ : Dict = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowerCamelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator A_ : Union[str, Any] = self.get_dummy_components() A_ : Any = StableDiffusionInstructPixaPixPipeline(**snake_case_ ) A_ : List[str] = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Tuple = self.get_dummy_inputs(snake_case_ ) A_ : Tuple = sd_pipe(**snake_case_ ).images A_ : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) A_ : int = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator A_ : Any = self.get_dummy_components() A_ : Tuple = StableDiffusionInstructPixaPixPipeline(**snake_case_ ) A_ : Tuple = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Dict = self.get_dummy_inputs(snake_case_ ) A_ : Tuple = 'french fries' A_ : List[str] = sd_pipe(**snake_case_ , negative_prompt=snake_case_ ) A_ : str = output.images A_ : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) A_ : Union[str, Any] = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator A_ : List[str] = self.get_dummy_components() A_ : List[str] = StableDiffusionInstructPixaPixPipeline(**snake_case_ ) A_ : Any = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Optional[Any] = self.get_dummy_inputs(snake_case_ ) A_ : Union[str, Any] = [inputs['prompt']] * 2 A_ : Tuple = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 A_ : int = torch.from_numpy(snake_case_ ).unsqueeze(0 ).to(snake_case_ ) A_ : Dict = image / 2 + 0.5 A_ : int = image.permute(0 , 3 , 1 , 2 ) A_ : Union[str, Any] = image.repeat(2 , 1 , 1 , 1 ) A_ : Tuple = sd_pipe(**snake_case_ ).images A_ : Optional[Any] = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) A_ : Any = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : str = 'cpu' # ensure determinism for the device-dependent torch.Generator A_ : List[str] = self.get_dummy_components() A_ : Dict = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) A_ : Dict = StableDiffusionInstructPixaPixPipeline(**snake_case_ ) A_ : Dict = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Optional[Any] = self.get_dummy_inputs(snake_case_ ) A_ : Optional[int] = sd_pipe(**snake_case_ ).images A_ : Tuple = image[0, -3:, -3:, -1] A_ : List[Any] = [round(snake_case_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(snake_case_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) A_ : List[str] = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[Any] = self.get_dummy_components() A_ : Any = StableDiffusionInstructPixaPixPipeline(**snake_case_ ) A_ : Optional[int] = VaeImageProcessor(do_resize=snake_case_ , do_normalize=snake_case_ ) A_ : List[Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Dict = pipe(**self.get_dummy_inputs_by_type(snake_case_ , input_image_type='pt' ) )[0] A_ : int = components['vae'] A_ : int = self.get_dummy_inputs_by_type(snake_case_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): A_ : int = vae.encode(inputs[image_param] ).latent_dist.mode() A_ : Union[str, Any] = pipe(**snake_case_ )[0] A_ : int = np.abs(out - out_latents_inputs ).max() self.assertLess(snake_case_ , 1E-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase_ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self , snake_case_=0 ): """simple docstring""" A_ : Optional[Any] = torch.manual_seed(snake_case_ ) A_ : List[Any] = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) A_ : Any = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : str = self.get_inputs() A_ : List[str] = pipe(**snake_case_ ).images A_ : List[Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) A_ : Union[str, Any] = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=snake_case_ ) A_ : List[str] = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : int = self.get_inputs() A_ : str = pipe(**snake_case_ ).images A_ : Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) A_ : Union[str, Any] = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=snake_case_ ) A_ : Dict = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : Optional[int] = self.get_inputs() A_ : Dict = pipe(**snake_case_ ).images A_ : str = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) A_ : Any = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Tuple = 0 def callback_fn(snake_case_ , snake_case_ , snake_case_ ) -> None: A_ : Union[str, Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: A_ : str = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) A_ : Dict = latents[0, -3:, -3:, -1] A_ : str = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: A_ : Dict = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) A_ : Union[str, Any] = latents[0, -3:, -3:, -1] A_ : Tuple = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 A_ : Dict = False A_ : List[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=snake_case_ , torch_dtype=torch.floataa ) A_ : List[Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : List[str] = self.get_inputs() pipe(**snake_case_ , callback=snake_case_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowerCamelCase_ ( self ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() A_ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=snake_case_ , torch_dtype=torch.floataa ) A_ : Union[str, Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() A_ : Optional[int] = self.get_inputs() A_ : Optional[Any] = pipe(**snake_case_ ) A_ : Any = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 A_ : Dict = inputs['image'].resize((5_0_4, 5_0_4) ) A_ : List[Any] = 'timbrooks/instruct-pix2pix' A_ : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( snake_case_ , safety_checker=snake_case_ , ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : Any = pipe(**snake_case_ ) A_ : int = output.images[0] A_ : List[str] = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) A_ : Optional[Any] = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
286
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , snake_case_ = 7_6_8 , ): """simple docstring""" super().__init__() A_ : Optional[int] = nn.Parameter(torch.zeros(1 , snake_case_ ) ) A_ : Optional[int] = nn.Parameter(torch.ones(1 , snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ = None , snake_case_ = None , ): """simple docstring""" A_ : str = nn.Parameter(self.mean.to(snake_case_ ).to(snake_case_ ) ) A_ : Optional[int] = nn.Parameter(self.std.to(snake_case_ ).to(snake_case_ ) ) return self def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Tuple = (embeds - self.mean) * 1.0 / self.std return embeds def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = (embeds * self.std) + self.mean return embeds
286
1
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""input_features""", """attention_mask"""] def __init__( self , snake_case_=8_0 , snake_case_=1_6_0_0_0 , snake_case_=0.0 , snake_case_=1_0 , snake_case_=2_5 , snake_case_="hamming_window" , snake_case_=3_27_68.0 , snake_case_=0.97 , snake_case_=1.0 , snake_case_=True , snake_case_=True , snake_case_=False , **snake_case_ , ): """simple docstring""" super().__init__(feature_size=snake_case_ , sampling_rate=snake_case_ , padding_value=snake_case_ , **snake_case_ ) A_ : List[Any] = feature_size A_ : str = sampling_rate A_ : List[str] = padding_value A_ : List[Any] = hop_length A_ : Optional[int] = win_length A_ : str = frame_signal_scale A_ : str = preemphasis_coeff A_ : Union[str, Any] = mel_floor A_ : Any = normalize_means A_ : Union[str, Any] = normalize_vars A_ : int = win_function A_ : Tuple = return_attention_mask A_ : Tuple = win_length * sampling_rate // 1_0_0_0 A_ : Tuple = hop_length * sampling_rate // 1_0_0_0 A_ : List[Any] = optimal_fft_length(self.sample_size ) A_ : Union[str, Any] = (self.n_fft // 2) + 1 def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.win_function == "hamming_window": A_ : List[str] = window_function(window_length=self.sample_size , name=self.win_function , periodic=snake_case_ ) else: A_ : Optional[int] = window_function(window_length=self.sample_size , name=self.win_function ) A_ : int = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) A_ : Any = spectrogram( one_waveform * self.frame_signal_scale , window=snake_case_ , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=snake_case_ , preemphasis=self.preemphasis_coeff , mel_filters=snake_case_ , mel_floor=self.mel_floor , log_mel='log' , ) return msfc_features.T def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" if self.normalize_means: A_ : str = x[:input_length].mean(axis=0 ) A_ : List[Any] = np.subtract(snake_case_ , snake_case_ ) if self.normalize_vars: A_ : str = x[:input_length].std(axis=0 ) A_ : Tuple = np.divide(snake_case_ , snake_case_ ) if input_length < x.shape[0]: A_ : List[Any] = padding_value # make sure array is in float32 A_ : Union[str, Any] = x.astype(np.floataa ) return x def lowerCamelCase_ ( self , snake_case_ , snake_case_ = None ): """simple docstring""" A_ : Any = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(snake_case_ , snake_case_ , self.padding_value ) for x, n in zip(snake_case_ , snake_case_ )] def __call__( self , snake_case_ , snake_case_ = False , snake_case_ = None , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , **snake_case_ , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" F""" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with""" F""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) A_ : Optional[int] = isinstance(snake_case_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) A_ : List[Any] = is_batched_numpy or ( isinstance(snake_case_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: A_ : int = [np.asarray(snake_case_ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(snake_case_ , np.ndarray ): A_ : Tuple = np.asarray(snake_case_ , dtype=np.floataa ) elif isinstance(snake_case_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): A_ : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: A_ : Dict = [raw_speech] # extract fbank features A_ : int = [self._extract_mfsc_features(snake_case_ ) for one_waveform in raw_speech] # convert into correct format for padding A_ : Any = BatchFeature({'input_features': features} ) A_ : Tuple = self.pad( snake_case_ , padding=snake_case_ , max_length=snake_case_ , truncation=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , **snake_case_ , ) # make sure list is in array format A_ : Optional[Any] = padded_inputs.get('input_features' ) if isinstance(input_features[0] , snake_case_ ): A_ : List[str] = [np.asarray(snake_case_ , dtype=np.floataa ) for feature in input_features] A_ : List[Any] = padded_inputs.get('attention_mask' ) if attention_mask is not None: A_ : Any = [np.asarray(snake_case_ , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: A_ : str = ( np.array(snake_case_ , dtype=np.intaa ) if self._get_padding_strategies(snake_case_ , max_length=snake_case_ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) A_ : Any = self.normalize( padded_inputs['input_features'] , attention_mask=snake_case_ ) if return_tensors is not None: A_ : Union[str, Any] = padded_inputs.convert_to_tensors(snake_case_ ) return padded_inputs
286
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCamelCase_ : Any = HfArgumentParser(InitializationArguments) lowerCamelCase_ : Union[str, Any] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowerCamelCase_ : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowerCamelCase_ : Tuple = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) lowerCamelCase_ : int = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowerCamelCase_ : Any = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
286
1
"""simple docstring""" lowerCamelCase_ : int = { 0: '0', 1: '1', 2: '2', 3: '3', 4: '4', 5: '5', 6: '6', 7: '7', 8: '8', 9: '9', 10: 'a', 11: 'b', 12: 'c', 13: 'd', 14: 'e', 15: 'f', } def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" assert type(_UpperCAmelCase ) in (int, float) and decimal == int(_UpperCAmelCase ) A_ : Any = int(_UpperCAmelCase ) A_ : str = '' A_ : List[Any] = False if decimal < 0: A_ : Optional[Any] = True decimal *= -1 while decimal > 0: A_ , A_ : List[Any] = divmod(_UpperCAmelCase , 16 ) A_ : Optional[int] = values[remainder] + hexadecimal A_ : Optional[int] = '0x' + hexadecimal if negative: A_ : List[str] = '-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser lowerCamelCase_ : Any = re.compile(r'\s+') def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return {"hash": hashlib.mda(re.sub(_UpperCAmelCase , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [len(_UpperCAmelCase ) for line in example['content'].splitlines()] return {"line_mean": np.mean(_UpperCAmelCase ), "line_max": max(_UpperCAmelCase )} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 ): """simple docstring""" A_ : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] A_ : List[str] = example['content'].splitlines() for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 , _UpperCAmelCase=0.05 ): """simple docstring""" A_ : Any = ['unit tests', 'test file', 'configuration file'] A_ : Dict = example['content'].splitlines() A_ : List[Any] = 0 A_ : str = 0 # first test for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test A_ : Tuple = example['content'].count('\n' ) A_ : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = ['def ', 'class ', 'for ', 'while '] A_ : Tuple = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=4 ): """simple docstring""" A_ : Union[str, Any] = example['content'].splitlines() A_ : Any = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = tokenizer(example['content'] , truncation=_UpperCAmelCase )['input_ids'] A_ : Dict = len(example['content'] ) / len(_UpperCAmelCase ) return {"ratio": ratio} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = {} results.update(get_hash(_UpperCAmelCase ) ) results.update(line_stats(_UpperCAmelCase ) ) results.update(alpha_stats(_UpperCAmelCase ) ) results.update(char_token_ratio(_UpperCAmelCase ) ) results.update(is_autogenerated(_UpperCAmelCase ) ) results.update(is_config_or_test(_UpperCAmelCase ) ) results.update(has_no_keywords(_UpperCAmelCase ) ) results.update(has_few_assignments(_UpperCAmelCase ) ) return results def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if not check_uniques(_UpperCAmelCase , _UpperCAmelCase ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'rb' ) as f_in: with gzip.open(str(_UpperCAmelCase ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) os.unlink(_UpperCAmelCase ) # Settings lowerCamelCase_ : Optional[int] = HfArgumentParser(PreprocessingArguments) lowerCamelCase_ : Optional[Any] = parser.parse_args() if args.num_workers is None: lowerCamelCase_ : int = multiprocessing.cpu_count() lowerCamelCase_ : Tuple = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset lowerCamelCase_ : Tuple = time.time() lowerCamelCase_ : Tuple = load_dataset(args.dataset_name, split='train') print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing lowerCamelCase_ : List[str] = time.time() lowerCamelCase_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes lowerCamelCase_ : int = set(ds.unique('hash')) lowerCamelCase_ : Union[str, Any] = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics lowerCamelCase_ : Optional[int] = time.time() lowerCamelCase_ : Tuple = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: lowerCamelCase_ : Union[str, Any] = time.time() lowerCamelCase_ , lowerCamelCase_ : str = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file lowerCamelCase_ : Tuple = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) lowerCamelCase_ : Optional[Any] = output_dir / 'data' data_dir.mkdir(exist_ok=True) lowerCamelCase_ : List[str] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): lowerCamelCase_ : Optional[int] = str(data_dir / F"file-{file_number+1:012}.json") lowerCamelCase_ : List[str] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = [1] A_ , A_ , A_ : Tuple = 0, 0, 0 A_ : Dict = ugly_nums[ia] * 2 A_ : Dict = ugly_nums[ia] * 3 A_ : int = ugly_nums[ia] * 5 for _ in range(1 , _UpperCAmelCase ): A_ : List[str] = min(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) ugly_nums.append(_UpperCAmelCase ) if next_num == next_a: ia += 1 A_ : Tuple = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 A_ : List[str] = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 A_ : Union[str, Any] = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(F"{ugly_numbers(2_00) = }")
286
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : Optional[Any] = [] # fmt: off # stem: rename_keys.append(('cls_token', 'vit.embeddings.cls_token') ) rename_keys.append(('pos_embed', 'vit.embeddings.position_embeddings') ) rename_keys.append(('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias') ) # backbone rename_keys.append(('patch_embed.backbone.stem.conv.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.bias', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A_ : List[str] = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) # fmt: on return rename_keys def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: A_ : List[str] = '' else: A_ : Dict = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ : str = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) A_ : List[Any] = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A_ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A_ : Tuple = in_proj_bias[: config.hidden_size] A_ : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A_ : Tuple = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = dct.pop(_UpperCAmelCase ) A_ : Optional[int] = val def UpperCAmelCase__ ( ): """simple docstring""" A_ : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' A_ : int = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : List[Any] = BitConfig( global_padding='same' , layer_type='bottleneck' , depths=(3, 4, 9) , out_features=['stage3'] , embedding_dynamic_padding=_UpperCAmelCase , ) A_ : Optional[int] = ViTHybridConfig(backbone_config=_UpperCAmelCase , image_size=384 , num_labels=1000 ) A_ : Union[str, Any] = False # load original model from timm A_ : List[Any] = timm.create_model(_UpperCAmelCase , pretrained=_UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ : Tuple = timm_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) A_ : Any = create_rename_keys(_UpperCAmelCase , _UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : Union[str, Any] = 'huggingface/label-files' A_ : Dict = 'imagenet-1k-id2label.json' A_ : List[str] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) A_ : str = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} A_ : Any = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": A_ : List[Any] = ViTHybridModel(_UpperCAmelCase ).eval() else: A_ : str = ViTHybridForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # create image processor A_ : Dict = create_transform(**resolve_data_config({} , model=_UpperCAmelCase ) ) A_ : List[str] = transform.transforms A_ : List[str] = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } A_ : Tuple = ViTHybridImageProcessor( do_resize=_UpperCAmelCase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_UpperCAmelCase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=_UpperCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) A_ : Optional[Any] = prepare_img() A_ : Any = transform(_UpperCAmelCase ).unsqueeze(0 ) A_ : Dict = processor(_UpperCAmelCase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase ) # verify logits with torch.no_grad(): A_ : List[Any] = model(_UpperCAmelCase ) A_ : List[str] = outputs.logits print('Predicted class:' , logits.argmax(-1 ).item() ) if base_model: A_ : Union[str, Any] = timm_model.forward_features(_UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_UpperCAmelCase , outputs.pooler_output , atol=1E-3 ) else: A_ : Tuple = timm_model(_UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase , outputs.logits , atol=1E-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(f"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: print(f"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(f"""ybelkada/{vit_name}""" ) processor.push_to_hub(f"""ybelkada/{vit_name}""" ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_r50_s16_384', type=str, help='Name of the hybrid ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) lowerCamelCase_ : List[str] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
286
1
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : int = len(_UpperCAmelCase ) A_ : int = len(_UpperCAmelCase ) A_ : int = ( first_str_length if first_str_length > second_str_length else second_str_length ) A_ : list = [] for char_count in range(_UpperCAmelCase ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(_UpperCAmelCase ) if __name__ == "__main__": print(alternative_string_arrange('AB', 'XYZ'), end=' ')
286
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" A_ : str = False if num < 0: A_ : Dict = True A_ : Union[str, Any] = -num A_ : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(_UpperCAmelCase ) for e in binary ) return "0b" + "".join(str(_UpperCAmelCase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowercase_ : str = StableDiffusionInpaintPipeline lowercase_ : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowercase_ : int = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowercase_ : int = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowercase_ : Optional[int] = frozenset([] ) def lowerCamelCase_ ( self ): """simple docstring""" torch.manual_seed(0 ) A_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=snake_case_ , ) A_ : Optional[Any] = PNDMScheduler(skip_prk_steps=snake_case_ ) torch.manual_seed(0 ) A_ : List[str] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) A_ : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='gelu' , projection_dim=5_1_2 , ) A_ : Any = CLIPTextModel(snake_case_ ) A_ : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) A_ : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowerCamelCase_ ( self , snake_case_ , snake_case_=0 ): """simple docstring""" A_ : Optional[int] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 )[0] A_ : List[str] = Image.fromarray(np.uinta(snake_case_ ) ).convert('RGB' ).resize((6_4, 6_4) ) A_ : int = Image.fromarray(np.uinta(image + 4 ) ).convert('RGB' ).resize((6_4, 6_4) ) if str(snake_case_ ).startswith('mps' ): A_ : List[str] = torch.manual_seed(snake_case_ ) else: A_ : Any = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) A_ : Any = { 'prompt': 'A painting of a squirrel eating a burger', 'image': init_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def lowerCamelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator A_ : Tuple = self.get_dummy_components() A_ : str = StableDiffusionInpaintPipeline(**snake_case_ ) A_ : str = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Dict = self.get_dummy_inputs(snake_case_ ) A_ : Dict = sd_pipe(**snake_case_ ).images A_ : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) A_ : Optional[Any] = np.array([0.47_27, 0.57_35, 0.39_41, 0.54_46, 0.59_26, 0.43_94, 0.50_62, 0.46_54, 0.44_76] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase_ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase_ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) A_ : Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) A_ : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench.npy' ) A_ : Optional[int] = 'stabilityai/stable-diffusion-2-inpainting' A_ : Any = StableDiffusionInpaintPipeline.from_pretrained(snake_case_ , safety_checker=snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : str = 'Face of a yellow cat, high resolution, sitting on a park bench' A_ : str = torch.manual_seed(0 ) A_ : Optional[int] = pipe( prompt=snake_case_ , image=snake_case_ , mask_image=snake_case_ , generator=snake_case_ , output_type='np' , ) A_ : Union[str, Any] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) A_ : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) A_ : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench_fp16.npy' ) A_ : Dict = 'stabilityai/stable-diffusion-2-inpainting' A_ : List[Any] = StableDiffusionInpaintPipeline.from_pretrained( snake_case_ , torch_dtype=torch.floataa , safety_checker=snake_case_ , ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() A_ : Union[str, Any] = 'Face of a yellow cat, high resolution, sitting on a park bench' A_ : Tuple = torch.manual_seed(0 ) A_ : Union[str, Any] = pipe( prompt=snake_case_ , image=snake_case_ , mask_image=snake_case_ , generator=snake_case_ , output_type='np' , ) A_ : Optional[Any] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def lowerCamelCase_ ( self ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() A_ : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) A_ : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) A_ : Union[str, Any] = 'stabilityai/stable-diffusion-2-inpainting' A_ : int = PNDMScheduler.from_pretrained(snake_case_ , subfolder='scheduler' ) A_ : Dict = StableDiffusionInpaintPipeline.from_pretrained( snake_case_ , safety_checker=snake_case_ , scheduler=snake_case_ , torch_dtype=torch.floataa , ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() A_ : Dict = 'Face of a yellow cat, high resolution, sitting on a park bench' A_ : Dict = torch.manual_seed(0 ) A_ : int = pipe( prompt=snake_case_ , image=snake_case_ , mask_image=snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type='np' , ) A_ : Optional[Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
286
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCamelCase_ : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Tuple = ['MLukeTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys lowerCamelCase_ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
1
"""simple docstring""" from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING lowerCamelCase_ : Union[str, Any] = logging.get_logger(__name__) @add_end_docstrings(UpperCAmelCase__ ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) requires_backends(self , 'vision' ) self.check_model_type(snake_case_ ) def __call__( self , snake_case_ , **snake_case_ ): """simple docstring""" return super().__call__(snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , **snake_case_ ): """simple docstring""" return {}, {}, {} def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Dict = load_image(snake_case_ ) A_ : Union[str, Any] = image.size A_ : Optional[int] = self.image_processor(images=snake_case_ , return_tensors=self.framework ) return model_inputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Optional[int] = self.model(**snake_case_ ) return model_outputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Tuple = model_outputs.predicted_depth A_ : Union[str, Any] = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode='bicubic' , align_corners=snake_case_ ) A_ : Any = prediction.squeeze().cpu().numpy() A_ : Union[str, Any] = (output * 2_5_5 / np.max(snake_case_ )).astype('uint8' ) A_ : List[str] = Image.fromarray(snake_case_ ) A_ : Optional[Any] = {} A_ : Dict = predicted_depth A_ : List[str] = depth return output_dict
286
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers lowerCamelCase_ : List[str] = [int(0.5 * n * (n + 1)) for n in range(1, 1_01)] def UpperCAmelCase__ ( ): """simple docstring""" A_ : Union[str, Any] = os.path.dirname(os.path.realpath(_UpperCAmelCase ) ) A_ : Tuple = os.path.join(_UpperCAmelCase , 'words.txt' ) A_ : List[Any] = '' with open(_UpperCAmelCase ) as f: A_ : int = f.readline() A_ : Optional[Any] = [word.strip('"' ) for word in words.strip('\r\n' ).split(',' )] A_ : Dict = [ word for word in [sum(ord(_UpperCAmelCase ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
286
1
"""simple docstring""" import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , ): """simple docstring""" if config_name_or_path is None: A_ : Optional[int] = 'facebook/rag-token-base' if model_type == 'rag_token' else 'facebook/rag-sequence-base' if generator_tokenizer_name_or_path is None: A_ : str = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: A_ : str = question_encoder_name_or_path A_ : int = RagTokenForGeneration if model_type == 'rag_token' else RagSequenceForGeneration # Save model. A_ : Optional[Any] = RagConfig.from_pretrained(_UpperCAmelCase ) A_ : str = AutoConfig.from_pretrained(_UpperCAmelCase ) A_ : Tuple = AutoConfig.from_pretrained(_UpperCAmelCase ) A_ : str = gen_config A_ : List[Any] = question_encoder_config A_ : Optional[int] = model_class.from_pretrained_question_encoder_generator( _UpperCAmelCase , _UpperCAmelCase , config=_UpperCAmelCase ) rag_model.save_pretrained(_UpperCAmelCase ) # Sanity check. model_class.from_pretrained(_UpperCAmelCase ) # Save tokenizers. A_ : List[str] = AutoTokenizer.from_pretrained(_UpperCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / 'generator_tokenizer/' ) A_ : Tuple = AutoTokenizer.from_pretrained(_UpperCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / 'question_encoder_tokenizer/' ) if __name__ == "__main__": lowerCamelCase_ : List[Any] = argparse.ArgumentParser() parser.add_argument( '--model_type', choices=['rag_sequence', 'rag_token'], required=True, type=str, help='RAG model type: rag_sequence, rag_token', ) parser.add_argument('--dest', type=str, required=True, help='Path to the output checkpoint directory.') parser.add_argument('--generator_name_or_path', type=str, required=True, help='Generator model identifier') parser.add_argument( '--question_encoder_name_or_path', type=str, required=True, help='Question encoder model identifier' ) parser.add_argument( '--generator_tokenizer_name_or_path', type=str, help='Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``', ) parser.add_argument( '--question_encoder_tokenizer_name_or_path', type=str, help='Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``', ) parser.add_argument( '--config_name_or_path', type=str, help=( 'Identifier of the model config to use, if not provided, resolves to a base config for a given' ' ``model_type``' ), ) lowerCamelCase_ : Optional[int] = parser.parse_args() lowerCamelCase_ : int = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
286
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ : List[str] = {'configuration_xlnet': ['XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : str = ['XLNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[str] = ['XLNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ 'XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLNetForMultipleChoice', 'XLNetForQuestionAnswering', 'XLNetForQuestionAnsweringSimple', 'XLNetForSequenceClassification', 'XLNetForTokenClassification', 'XLNetLMHeadModel', 'XLNetModel', 'XLNetPreTrainedModel', 'load_tf_weights_in_xlnet', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Union[str, Any] = [ 'TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLNetForMultipleChoice', 'TFXLNetForQuestionAnsweringSimple', 'TFXLNetForSequenceClassification', 'TFXLNetForTokenClassification', 'TFXLNetLMHeadModel', 'TFXLNetMainLayer', 'TFXLNetModel', 'TFXLNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase_ : str = {'configuration_wavlm': ['WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WavLMConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Any = [ 'WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'WavLMForAudioFrameClassification', 'WavLMForCTC', 'WavLMForSequenceClassification', 'WavLMForXVector', 'WavLMModel', 'WavLMPreTrainedModel', ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys lowerCamelCase_ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
"""simple docstring""" import torch from diffusers import DiffusionPipeline class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" super().__init__() self.register_modules(unet=snake_case_ , scheduler=snake_case_ ) def __call__( self ): """simple docstring""" A_ : Optional[Any] = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) A_ : List[str] = 1 A_ : List[str] = self.unet(snake_case_ , snake_case_ ).sample A_ : Optional[int] = self.scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample A_ : List[Any] = scheduler_output - scheduler_output + torch.ones_like(snake_case_ ) return result
286
1
"""simple docstring""" from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time lowerCamelCase_ : List[Any] = Lock() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(_UpperCAmelCase ) process_lock.release() # receive your right neighbor's value process_lock.acquire() A_ : Dict = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left A_ : str = min(_UpperCAmelCase , _UpperCAmelCase ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(_UpperCAmelCase ) process_lock.release() # receive your left neighbor's value process_lock.acquire() A_ : int = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right A_ : str = max(_UpperCAmelCase , _UpperCAmelCase ) # after all swaps are performed, send the values back to main result_pipe[1].send(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [] A_ : List[Any] = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop A_ : int = Pipe() A_ : Optional[Any] = Pipe() process_array_.append( Process( target=_UpperCAmelCase , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) A_ : str = temp_rs A_ : Optional[int] = temp_rr for i in range(1 , len(_UpperCAmelCase ) - 1 ): A_ : Tuple = Pipe() A_ : Dict = Pipe() process_array_.append( Process( target=_UpperCAmelCase , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) A_ : str = temp_rs A_ : Optional[Any] = temp_rr process_array_.append( Process( target=_UpperCAmelCase , args=( len(_UpperCAmelCase ) - 1, arr[len(_UpperCAmelCase ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(_UpperCAmelCase ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(_UpperCAmelCase ) ): A_ : str = result_pipe[p][0].recv() process_array_[p].join() return arr def UpperCAmelCase__ ( ): """simple docstring""" A_ : List[Any] = list(range(10 , 0 , -1 ) ) print('Initial List' ) print(*_UpperCAmelCase ) A_ : Optional[Any] = odd_even_transposition(_UpperCAmelCase ) print('Sorted List\n' ) print(*_UpperCAmelCase ) if __name__ == "__main__": main()
286
"""simple docstring""" from heapq import heappop, heappush import numpy as np def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" A_ , A_ : List[str] = grid.shape A_ : Optional[int] = [-1, 1, 0, 0] A_ : str = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] A_ , A_ : List[Any] = [(0, source)], set() A_ : Optional[Any] = np.full((rows, cols) , np.inf ) A_ : int = 0 A_ : Optional[int] = np.empty((rows, cols) , dtype=_UpperCAmelCase ) A_ : Optional[int] = None while queue: ((A_) , (A_)) : str = heappop(_UpperCAmelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: A_ : int = [] while (x, y) != source: path.append((x, y) ) A_ , A_ : List[Any] = predecessors[x, y] path.append(_UpperCAmelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(_UpperCAmelCase ) ): A_ , A_ : Tuple = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: A_ : Union[str, Any] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(_UpperCAmelCase , (dist + 1, (nx, ny)) ) A_ : Optional[Any] = dist + 1 A_ : Optional[Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Optional[Any] = arr.split(',' ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = [int(self.array[0] )] * len(self.array ) A_ : Optional[Any] = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): A_ : Optional[int] = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) A_ : Union[str, Any] = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": lowerCamelCase_ : str = input('please input some numbers:') lowerCamelCase_ : List[Any] = SubArray(whole_array) lowerCamelCase_ : Tuple = array.solve_sub_array() print(('the results is:', re))
286
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase_ : Optional[Any] = { 'huggingface/informer-tourism-monthly': ( 'https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json' ), # See all Informer models at https://huggingface.co/models?filter=informer } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """informer""" lowercase_ : str = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self , snake_case_ = None , snake_case_ = None , snake_case_ = "student_t" , snake_case_ = "nll" , snake_case_ = 1 , snake_case_ = None , snake_case_ = "mean" , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = 6_4 , snake_case_ = 3_2 , snake_case_ = 3_2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = True , snake_case_ = "gelu" , snake_case_ = 0.05 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 1_0_0 , snake_case_ = 0.02 , snake_case_=True , snake_case_ = "prob" , snake_case_ = 5 , snake_case_ = True , **snake_case_ , ): """simple docstring""" A_ : str = prediction_length A_ : List[Any] = context_length or prediction_length A_ : str = distribution_output A_ : Dict = loss A_ : Any = input_size A_ : Union[str, Any] = num_time_features A_ : Optional[Any] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] A_ : List[Any] = scaling A_ : Tuple = num_dynamic_real_features A_ : Any = num_static_real_features A_ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) A_ : Optional[int] = cardinality else: A_ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) A_ : Any = embedding_dimension else: A_ : Optional[Any] = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] A_ : int = num_parallel_samples # Transformer architecture configuration A_ : str = input_size * len(self.lags_sequence ) + self._number_of_features A_ : List[Any] = d_model A_ : Dict = encoder_attention_heads A_ : Dict = decoder_attention_heads A_ : List[Any] = encoder_ffn_dim A_ : Union[str, Any] = decoder_ffn_dim A_ : int = encoder_layers A_ : Any = decoder_layers A_ : List[Any] = dropout A_ : str = attention_dropout A_ : Tuple = activation_dropout A_ : List[str] = encoder_layerdrop A_ : List[str] = decoder_layerdrop A_ : str = activation_function A_ : Optional[int] = init_std A_ : List[Any] = use_cache # Informer A_ : Tuple = attention_type A_ : List[Any] = sampling_factor A_ : Optional[int] = distil super().__init__(is_encoder_decoder=snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
286
1
"""simple docstring""" import argparse from collections import defaultdict def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Optional[int] = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_UpperCAmelCase , 'r' ) as f: A_ : Any = f.readlines() A_ : int = f"""class {class_name}(""" A_ : List[str] = f"""{4 * " "}def {test_name}(""" A_ : Optional[int] = f"""{8 * " "}{correct_line.split()[0]}""" A_ : int = f"""{16 * " "}{correct_line.split()[0]}""" A_ : Optional[int] = False A_ : Tuple = False A_ : Any = False A_ : Tuple = False A_ : Any = 0 A_ : int = 0 A_ : int = [] for line in lines: if line.startswith(_UpperCAmelCase ): A_ : Optional[int] = True elif in_class and line.startswith(_UpperCAmelCase ): A_ : Optional[int] = True elif in_class and in_func and (line.startswith(_UpperCAmelCase ) or line.startswith(_UpperCAmelCase )): A_ : Union[str, Any] = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: A_ : List[Any] = True if in_class and in_func and in_line: if ")" not in line: continue else: A_ : Tuple = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) A_ : Dict = False else: new_lines.append(_UpperCAmelCase ) with open(_UpperCAmelCase , 'w' ) as f: for line in new_lines: f.write(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=None ): """simple docstring""" if fail is not None: with open(_UpperCAmelCase , 'r' ) as f: A_ : str = {l.strip() for l in f.readlines()} else: A_ : List[str] = None with open(_UpperCAmelCase , 'r' ) as f: A_ : int = f.readlines() A_ : Optional[Any] = defaultdict(_UpperCAmelCase ) for line in correct_lines: A_ , A_ , A_ , A_ : Any = line.split(';' ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) lowerCamelCase_ : Optional[Any] = parser.parse_args() main(args.correct_filename, args.fail_filename)
286
"""simple docstring""" import os def UpperCAmelCase__ ( ): """simple docstring""" A_ : Any = os.path.join(os.path.dirname(_UpperCAmelCase ) , 'num.txt' ) with open(_UpperCAmelCase ) as file_hand: return str(sum(int(_UpperCAmelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
286
1
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCamelCase_ : Dict = NewType('DataClass', Any) lowerCamelCase_ : Any = NewType('DataClassType', Any) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Tuple = {str(_UpperCAmelCase ): choice for choice in choices} return lambda _UpperCAmelCase : str_to_choice.get(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( *, _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = dataclasses.MISSING , _UpperCAmelCase = dataclasses.MISSING , _UpperCAmelCase = None , **_UpperCAmelCase , ): """simple docstring""" if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls A_ : Optional[Any] = {} if aliases is not None: A_ : Tuple = aliases if help is not None: A_ : List[Any] = help return dataclasses.field(metadata=_UpperCAmelCase , default=_UpperCAmelCase , default_factory=_UpperCAmelCase , **_UpperCAmelCase ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Iterable[DataClassType] def __init__( self , snake_case_ , **snake_case_ ): """simple docstring""" if "formatter_class" not in kwargs: A_ : Union[str, Any] = ArgumentDefaultsHelpFormatter super().__init__(**snake_case_ ) if dataclasses.is_dataclass(snake_case_ ): A_ : List[str] = [dataclass_types] A_ : Any = list(snake_case_ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(snake_case_ ) @staticmethod def lowerCamelCase_ ( snake_case_ , snake_case_ ): """simple docstring""" A_ : List[Any] = F"""--{field.name}""" A_ : Optional[int] = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , snake_case_ ): raise RuntimeError( 'Unresolved type detected, which should have been done with the help of ' '`typing.get_type_hints` method by default' ) A_ : int = kwargs.pop('aliases' , [] ) if isinstance(snake_case_ , snake_case_ ): A_ : List[str] = [aliases] A_ : List[str] = getattr(field.type , '__origin__' , field.type ) if origin_type is Union or (hasattr(snake_case_ , 'UnionType' ) and isinstance(snake_case_ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(snake_case_ ) not in field.type.__args__ ): raise ValueError( 'Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because' ' the argument parser only supports one type per argument.' F""" Problem encountered in field '{field.name}'.""" ) if type(snake_case_ ) not in field.type.__args__: # filter `str` in Union A_ : Optional[int] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] A_ : List[Any] = getattr(field.type , '__origin__' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) A_ : int = ( field.type.__args__[0] if isinstance(snake_case_ , field.type.__args__[1] ) else field.type.__args__[1] ) A_ : Optional[int] = getattr(field.type , '__origin__' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) A_ : Union[str, Any] = {} if origin_type is Literal or (isinstance(field.type , snake_case_ ) and issubclass(field.type , snake_case_ )): if origin_type is Literal: A_ : Optional[int] = field.type.__args__ else: A_ : Optional[Any] = [x.value for x in field.type] A_ : List[str] = make_choice_type_function(kwargs['choices'] ) if field.default is not dataclasses.MISSING: A_ : Tuple = field.default else: A_ : Any = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument A_ : Dict = copy(snake_case_ ) # Hack because type=bool in argparse does not behave as we want. A_ : List[Any] = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. A_ : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way A_ : int = default # This tells argparse we accept 0 or 1 value after --field_name A_ : int = '?' # This is the value that will get picked if we do --field_name (without value) A_ : Dict = True elif isclass(snake_case_ ) and issubclass(snake_case_ , snake_case_ ): A_ : int = field.type.__args__[0] A_ : Tuple = '+' if field.default_factory is not dataclasses.MISSING: A_ : Dict = field.default_factory() elif field.default is dataclasses.MISSING: A_ : Dict = True else: A_ : List[str] = field.type if field.default is not dataclasses.MISSING: A_ : Union[str, Any] = field.default elif field.default_factory is not dataclasses.MISSING: A_ : Tuple = field.default_factory() else: A_ : Tuple = True parser.add_argument(snake_case_ , *snake_case_ , **snake_case_ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): A_ : List[str] = False parser.add_argument(F"""--no_{field.name}""" , action='store_false' , dest=field.name , **snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if hasattr(snake_case_ , '_argument_group_name' ): A_ : List[str] = self.add_argument_group(dtype._argument_group_name ) else: A_ : Union[str, Any] = self try: A_ : Dict[str, type] = get_type_hints(snake_case_ ) except NameError: raise RuntimeError( F"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ 'removing line of `from __future__ import annotations` which opts in Postponed ' 'Evaluation of Annotations (PEP 563)' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(snake_case_ ): A_ : Optional[int] = '.'.join(map(snake_case_ , sys.version_info[:3] ) ) raise RuntimeError( F"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ 'line of `from __future__ import annotations` which opts in union types as ' '`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ' 'support Python versions that lower than 3.10, you need to use ' '`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ' '`X | None`.' ) from ex raise for field in dataclasses.fields(snake_case_ ): if not field.init: continue A_ : str = type_hints[field.name] self._parse_dataclass_field(snake_case_ , snake_case_ ) def lowerCamelCase_ ( self , snake_case_=None , snake_case_=False , snake_case_=True , snake_case_=None , snake_case_=None , ): """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): A_ : Dict = [] if args_filename: args_files.append(Path(snake_case_ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('.args' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values A_ : Optional[Any] = ArgumentParser() args_file_parser.add_argument(snake_case_ , type=snake_case_ , action='append' ) # Use only remaining args for further parsing (remove the args_file_flag) A_ , A_ : List[Any] = args_file_parser.parse_known_args(args=snake_case_ ) A_ : Any = vars(snake_case_ ).get(args_file_flag.lstrip('-' ) , snake_case_ ) if cmd_args_file_paths: args_files.extend([Path(snake_case_ ) for p in cmd_args_file_paths] ) A_ : List[Any] = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last A_ : Tuple = file_args + args if args is not None else file_args + sys.argv[1:] A_ , A_ : str = self.parse_known_args(args=snake_case_ ) A_ : Optional[int] = [] for dtype in self.dataclass_types: A_ : int = {f.name for f in dataclasses.fields(snake_case_ ) if f.init} A_ : List[str] = {k: v for k, v in vars(snake_case_ ).items() if k in keys} for k in keys: delattr(snake_case_ , snake_case_ ) A_ : List[Any] = dtype(**snake_case_ ) outputs.append(snake_case_ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(snake_case_ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def lowerCamelCase_ ( self , snake_case_ , snake_case_ = False ): """simple docstring""" A_ : Optional[int] = set(args.keys() ) A_ : int = [] for dtype in self.dataclass_types: A_ : Dict = {f.name for f in dataclasses.fields(snake_case_ ) if f.init} A_ : int = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) A_ : Union[str, Any] = dtype(**snake_case_ ) outputs.append(snake_case_ ) if not allow_extra_keys and unused_keys: raise ValueError(F"""Some keys are not used by the HfArgumentParser: {sorted(snake_case_ )}""" ) return tuple(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ = False ): """simple docstring""" with open(Path(snake_case_ ) , encoding='utf-8' ) as open_json_file: A_ : Optional[Any] = json.loads(open_json_file.read() ) A_ : List[str] = self.parse_dict(snake_case_ , allow_extra_keys=snake_case_ ) return tuple(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ = False ): """simple docstring""" A_ : str = self.parse_dict(yaml.safe_load(Path(snake_case_ ).read_text() ) , allow_extra_keys=snake_case_ ) return tuple(snake_case_ )
286
"""simple docstring""" import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger lowerCamelCase_ : Dict = get_logger(__name__) lowerCamelCase_ : List[str] = r'\n Args:\n input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam\n search or log softmax for each vocabulary token when using beam search\n kwargs (`Dict[str, Any]`, *optional*):\n Additional logits processor specific kwargs.\n\n Return:\n `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores.\n\n' class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ): """simple docstring""" for processor in self: A_ : Tuple = inspect.signature(processor.__call__ ).parameters if len(snake_case_ ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( F"""Make sure that all the required parameters: {list(function_args.keys() )} for """ F"""{processor.__class__} are passed to the logits processor.""" ) A_ : Tuple = processor(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) else: A_ : Optional[Any] = processor(snake_case_ , snake_case_ , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or not (temperature > 0): raise ValueError(F"""`temperature` has to be a strictly positive float, but is {temperature}""" ) A_ : Optional[int] = temperature def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = scores / self.temperature return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or (top_p < 0 or top_p > 1.0): raise ValueError(F"""`top_p` has to be a float > 0 and < 1, but is {top_p}""" ) if not isinstance(snake_case_ , snake_case_ ) or (min_tokens_to_keep < 1): raise ValueError(F"""`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}""" ) A_ : str = top_p A_ : Union[str, Any] = filter_value A_ : int = min_tokens_to_keep def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : Tuple = lax.top_k(snake_case_ , scores.shape[-1] ) A_ : List[Any] = jnp.full_like(snake_case_ , self.filter_value ) A_ : List[str] = jax.nn.softmax(snake_case_ , axis=-1 ).cumsum(axis=-1 ) A_ : Optional[int] = cumulative_probs < self.top_p # include the token that is higher than top_p as well A_ : Union[str, Any] = jnp.roll(snake_case_ , 1 ) score_mask |= score_mask.at[:, 0].set(snake_case_ ) # min tokens to keep A_ : int = score_mask.at[:, : self.min_tokens_to_keep].set(snake_case_ ) A_ : Optional[Any] = jnp.where(snake_case_ , snake_case_ , snake_case_ ) A_ : List[Any] = jax.lax.sort_key_val(snake_case_ , snake_case_ )[-1] return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or top_k <= 0: raise ValueError(F"""`top_k` has to be a strictly positive integer, but is {top_k}""" ) A_ : str = max(snake_case_ , snake_case_ ) A_ : Union[str, Any] = filter_value def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : int = scores.shape A_ : Tuple = jnp.full(batch_size * vocab_size , self.filter_value ) A_ : Union[str, Any] = min(self.top_k , scores.shape[-1] ) # Safety check A_ , A_ : Dict = lax.top_k(snake_case_ , snake_case_ ) A_ : Optional[int] = jnp.broadcast_to((jnp.arange(snake_case_ ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() A_ : int = topk_scores.flatten() A_ : Any = topk_indices.flatten() + shift A_ : List[str] = next_scores_flat.at[topk_indices_flat].set(snake_case_ ) A_ : Union[str, Any] = next_scores_flat.reshape(snake_case_ , snake_case_ ) return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = bos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Union[str, Any] = 1 - jnp.bool_(cur_len - 1 ) A_ : str = jnp.where(snake_case_ , new_scores.at[:, self.bos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = max_length A_ : Optional[int] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Dict = 1 - jnp.bool_(cur_len - self.max_length + 1 ) A_ : Dict = jnp.where(snake_case_ , new_scores.at[:, self.eos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or min_length < 0: raise ValueError(F"""`min_length` has to be a positive integer, but is {min_length}""" ) if not isinstance(snake_case_ , snake_case_ ) or eos_token_id < 0: raise ValueError(F"""`eos_token_id` has to be a positive integer, but is {eos_token_id}""" ) A_ : Any = min_length A_ : List[Any] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) A_ : Optional[Any] = jnp.where(snake_case_ , scores.at[:, self.eos_token_id].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) A_ : Tuple = begin_index def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = 1 - jnp.bool_(cur_len - self.begin_index ) A_ : int = jnp.where(snake_case_ , scores.at[:, self.begin_suppress_tokens].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = scores.at[..., self.suppress_tokens].set(-float('inf' ) ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Any = dict(snake_case_ ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. A_ : Tuple = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: A_ : Tuple = force_token_array.at[index].set(snake_case_ ) A_ : Any = jnp.intaa(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" def _force_token(snake_case_ ): A_ : List[Any] = scores.shape[0] A_ : Any = self.force_token_array[generation_idx] A_ : Tuple = jnp.ones_like(snake_case_ , dtype=scores.dtype ) * -float('inf' ) A_ : List[Any] = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) A_ : int = lax.dynamic_update_slice(snake_case_ , snake_case_ , (0, current_token) ) return new_scores A_ : int = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(snake_case_ ) , lambda: scores , ) , ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = generate_config.eos_token_id A_ : Optional[int] = generate_config.no_timestamps_token_id A_ : List[str] = generate_config.no_timestamps_token_id + 1 A_ : Any = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(snake_case_ , 'max_initial_timestamp_index' ): A_ : List[Any] = generate_config.max_initial_timestamp_index else: A_ : Any = model_config.vocab_size if self.max_initial_timestamp_index is None: A_ : Optional[Any] = model_config.vocab_size def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[str] = scores.at[:, self.no_timestamps_token_id].set(-float('inf' ) ) def handle_pairs(snake_case_ , snake_case_ ): A_ : Any = jnp.where((cur_len - self.begin_index) >= 1 , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , snake_case_ , ) A_ : Tuple = jnp.where((cur_len - self.begin_index) < 2 , snake_case_ , snake_case_ ) A_ : Any = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , snake_case_ , snake_case_ , ) return jnp.where( snake_case_ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float('inf' ) ) , scores_k.at[: self.eos_token_id].set(-float('inf' ) ) , ) , snake_case_ , ) A_ : Tuple = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) A_ : Optional[Any] = jnp.where(cur_len == self.begin_index , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , snake_case_ , ) A_ : int = self.timestamp_begin + self.max_initial_timestamp_index A_ : List[Any] = jnp.where( snake_case_ , scores.at[:, last_allowed + 1 :].set(-float('inf' ) ) , snake_case_ , ) # if sum of probability over timestamps is above any other token, sample timestamp A_ : Any = jax.nn.log_softmax(snake_case_ , axis=-1 ) def handle_cumulative_probs(snake_case_ , snake_case_ ): A_ : Dict = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) A_ : Optional[Any] = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float('inf' ) ) , snake_case_ , ) A_ : Union[str, Any] = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) return scores
286
1
"""simple docstring""" from collections.abc import Sequence from queue import Queue class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None , snake_case_=None ): """simple docstring""" A_ : Dict = start A_ : Optional[Any] = end A_ : Any = val A_ : Union[str, Any] = (start + end) // 2 A_ : List[str] = left A_ : Any = right def __repr__( self ): """simple docstring""" return F"""SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})""" class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = collection A_ : Any = function if self.collection: A_ : Dict = self._build_tree(0 , len(snake_case_ ) - 1 ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" self._update_tree(self.root , snake_case_ , snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" return self._query_range(self.root , snake_case_ , snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if start == end: return SegmentTreeNode(snake_case_ , snake_case_ , self.collection[start] ) A_ : int = (start + end) // 2 A_ : Dict = self._build_tree(snake_case_ , snake_case_ ) A_ : Any = self._build_tree(mid + 1 , snake_case_ ) return SegmentTreeNode(snake_case_ , snake_case_ , self.fn(left.val , right.val ) , snake_case_ , snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" if node.start == i and node.end == i: A_ : List[Any] = val return if i <= node.mid: self._update_tree(node.left , snake_case_ , snake_case_ ) else: self._update_tree(node.right , snake_case_ , snake_case_ ) A_ : Optional[Any] = self.fn(node.left.val , node.right.val ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , snake_case_ , snake_case_ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , snake_case_ , node.mid ) , self._query_range(node.right , node.mid + 1 , snake_case_ ) , ) else: # range in right child tree return self._query_range(node.right , snake_case_ , snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" if self.root is not None: A_ : Optional[int] = Queue() queue.put(self.root ) while not queue.empty(): A_ : Tuple = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('*' * 50) lowerCamelCase_ : Dict = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
286
"""simple docstring""" import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = R'\w+[.]\d+' A_ : int = re.findall(_UpperCAmelCase , _UpperCAmelCase ) for pat in pats: A_ : Optional[int] = key.replace(_UpperCAmelCase , '_'.join(pat.split('.' ) ) ) return key def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): A_ : Union[str, Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: A_ : List[str] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: A_ : Optional[Any] = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer A_ : int = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: A_ : str = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer A_ : Optional[Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": A_ : Optional[Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight A_ : Tuple = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias A_ : Optional[int] = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=42 ): """simple docstring""" A_ : int = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params A_ : Union[str, Any] = flax_model.init_weights(PRNGKey(_UpperCAmelCase ) ) A_ : Optional[Any] = flatten_dict(_UpperCAmelCase ) A_ : Tuple = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): A_ : Any = rename_key(_UpperCAmelCase ) A_ : List[str] = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters A_ , A_ : Union[str, Any] = rename_key_and_reshape_tensor(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown A_ : str = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase )
286
1
"""simple docstring""" import math def UpperCAmelCase__ ( ): """simple docstring""" A_ : List[str] = input('Enter message: ' ) A_ : str = int(input(f"""Enter key [2-{len(_UpperCAmelCase ) - 1}]: """ ) ) A_ : Optional[int] = input('Encryption/Decryption [e/d]: ' ) if mode.lower().startswith('e' ): A_ : Any = encrypt_message(_UpperCAmelCase , _UpperCAmelCase ) elif mode.lower().startswith('d' ): A_ : int = decrypt_message(_UpperCAmelCase , _UpperCAmelCase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f"""Output:\n{text + "|"}""" ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [''] * key for col in range(_UpperCAmelCase ): A_ : Optional[Any] = col while pointer < len(_UpperCAmelCase ): cipher_text[col] += message[pointer] pointer += key return "".join(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : str = math.ceil(len(_UpperCAmelCase ) / key ) A_ : Optional[Any] = key A_ : Any = (num_cols * num_rows) - len(_UpperCAmelCase ) A_ : Any = [''] * num_cols A_ : Dict = 0 A_ : str = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): A_ : str = 0 row += 1 return "".join(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
286
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = CustomTokenizer pass
286
1
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
"""simple docstring""" from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch lowerCamelCase_ : str = logging.get_logger(__name__) @add_end_docstrings( UpperCAmelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.framework == "tf": A_ : str = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": A_ : List[str] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ) else: raise ValueError('Unsupported framework' ) return masked_index def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = self.get_masked_index(snake_case_ ) A_ : str = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['input_ids'][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_=None , **snake_case_ ): """simple docstring""" if return_tensors is None: A_ : Any = self.framework A_ : Dict = self.tokenizer(snake_case_ , return_tensors=snake_case_ ) self.ensure_exactly_one_mask_token(snake_case_ ) return model_inputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Dict = self.model(**snake_case_ ) A_ : Optional[int] = model_inputs['input_ids'] return model_outputs def lowerCamelCase_ ( self , snake_case_ , snake_case_=5 , snake_case_=None ): """simple docstring""" if target_ids is not None and target_ids.shape[0] < top_k: A_ : str = target_ids.shape[0] A_ : Optional[Any] = model_outputs['input_ids'][0] A_ : List[Any] = model_outputs['logits'] if self.framework == "tf": A_ : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] A_ : Union[str, Any] = outputs.numpy() A_ : Optional[int] = outputs[0, masked_index, :] A_ : Optional[Any] = stable_softmax(snake_case_ , axis=-1 ) if target_ids is not None: A_ : Union[str, Any] = tf.gather_nd(tf.squeeze(snake_case_ , 0 ) , target_ids.reshape(-1 , 1 ) ) A_ : Optional[int] = tf.expand_dims(snake_case_ , 0 ) A_ : Any = tf.math.top_k(snake_case_ , k=snake_case_ ) A_ , A_ : str = topk.values.numpy(), topk.indices.numpy() else: A_ : int = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample A_ : Tuple = outputs[0, masked_index, :] A_ : List[str] = logits.softmax(dim=-1 ) if target_ids is not None: A_ : str = probs[..., target_ids] A_ , A_ : List[str] = probs.topk(snake_case_ ) A_ : List[Any] = [] A_ : int = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): A_ : str = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place A_ : Union[str, Any] = input_ids.numpy().copy() if target_ids is not None: A_ : str = target_ids[p].tolist() A_ : Union[str, Any] = p # Filter padding out: A_ : Any = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back A_ : Any = self.tokenizer.decode(snake_case_ , skip_special_tokens=snake_case_ ) A_ : Any = {'score': v, 'token': p, 'token_str': self.tokenizer.decode([p] ), 'sequence': sequence} row.append(snake_case_ ) result.append(snake_case_ ) if single_mask: return result[0] return result def lowerCamelCase_ ( self , snake_case_ , snake_case_=None ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : List[str] = [targets] try: A_ : Optional[int] = self.tokenizer.get_vocab() except Exception: A_ : int = {} A_ : Tuple = [] for target in targets: A_ : int = vocab.get(snake_case_ , snake_case_ ) if id_ is None: A_ : Tuple = self.tokenizer( snake_case_ , add_special_tokens=snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , max_length=1 , truncation=snake_case_ , )['input_ids'] if len(snake_case_ ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ 'We cannot replace it with anything meaningful, ignoring it' ) continue A_ : str = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) A_ : Tuple = list(set(snake_case_ ) ) if len(snake_case_ ) == 0: raise ValueError('At least one target must be provided when passed.' ) A_ : Optional[Any] = np.array(snake_case_ ) return target_ids def lowerCamelCase_ ( self , snake_case_=None , snake_case_=None ): """simple docstring""" A_ : List[str] = {} if targets is not None: A_ : Any = self.get_target_ids(snake_case_ , snake_case_ ) A_ : Optional[Any] = target_ids if top_k is not None: A_ : int = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , 'The tokenizer does not define a `mask_token`.' ) return {}, {}, postprocess_params def __call__( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : List[str] = super().__call__(snake_case_ , **snake_case_ ) if isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1: return outputs[0] return outputs
286
1
"""simple docstring""" import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging lowerCamelCase_ : Any = ( 'https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py' ) lowerCamelCase_ : Any = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase__ ( ): """simple docstring""" A_ : str = 'https://pypi.org/pypi/diffusers/json' A_ : List[str] = json.loads(request.urlopen(_UpperCAmelCase ).read() )['releases'].keys() return sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : version.Version(_UpperCAmelCase ) ) def UpperCAmelCase__ ( ): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(_UpperCAmelCase ) os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) A_ : Any = Path(_UpperCAmelCase ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" init_hf_modules() A_ : Dict = Path(_UpperCAmelCase ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) A_ : Tuple = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'r' , encoding='utf-8' ) as f: A_ : Optional[Any] = f.read() # Imports of the form `import .xxx` A_ : Optional[int] = re.findall('^\s*import\s+\.(\S+)\s*$' , _UpperCAmelCase , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , _UpperCAmelCase , flags=re.MULTILINE ) # Unique-ify return list(set(_UpperCAmelCase ) ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = False A_ : Dict = [module_file] A_ : Dict = [] # Let's recurse through all relative imports while not no_change: A_ : int = [] for f in files_to_check: new_imports.extend(get_relative_imports(_UpperCAmelCase ) ) A_ : int = Path(_UpperCAmelCase ).parent A_ : Tuple = [str(module_path / m ) for m in new_imports] A_ : str = [f for f in new_import_files if f not in all_relative_imports] A_ : int = [f"""{f}.py""" for f in new_import_files] A_ : Optional[Any] = len(_UpperCAmelCase ) == 0 all_relative_imports.extend(_UpperCAmelCase ) return all_relative_imports def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'r' , encoding='utf-8' ) as f: A_ : Any = f.read() # Imports of the form `import xxx` A_ : Any = re.findall('^\s*import\s+(\S+)\s*$' , _UpperCAmelCase , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , _UpperCAmelCase , flags=re.MULTILINE ) # Only keep the top-level module A_ : Any = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all A_ : List[Any] = list(set(_UpperCAmelCase ) ) A_ : Any = [] for imp in imports: try: importlib.import_module(_UpperCAmelCase ) except ImportError: missing_packages.append(_UpperCAmelCase ) if len(_UpperCAmelCase ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{", ".join(_UpperCAmelCase )}. Run `pip install {" ".join(_UpperCAmelCase )}`""" ) return get_relative_imports(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = module_path.replace(os.path.sep , '.' ) A_ : int = importlib.import_module(_UpperCAmelCase ) if class_name is None: return find_pipeline_class(_UpperCAmelCase ) return getattr(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" from ..pipelines import DiffusionPipeline A_ : int = dict(inspect.getmembers(_UpperCAmelCase , inspect.isclass ) ) A_ : List[str] = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , _UpperCAmelCase ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) A_ : Union[str, Any] = cls return pipeline_class def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = False , ): """simple docstring""" A_ : int = str(_UpperCAmelCase ) A_ : Optional[int] = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if os.path.isfile(_UpperCAmelCase ): A_ : List[str] = module_file_or_url A_ : Union[str, Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: A_ : Union[str, Any] = get_diffusers_versions() # cut ".dev0" A_ : Optional[Any] = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: A_ : List[Any] = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: A_ : Optional[int] = f"""v{revision}""" elif revision == "main": A_ : Union[str, Any] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {", ".join(available_versions + ["main"] )}.""" ) # community pipeline on GitHub A_ : List[Any] = COMMUNITY_PIPELINES_URL.format(revision=_UpperCAmelCase , pipeline=_UpperCAmelCase ) try: A_ : Dict = cached_download( _UpperCAmelCase , cache_dir=_UpperCAmelCase , force_download=_UpperCAmelCase , proxies=_UpperCAmelCase , resume_download=_UpperCAmelCase , local_files_only=_UpperCAmelCase , use_auth_token=_UpperCAmelCase , ) A_ : Optional[int] = 'git' A_ : Optional[Any] = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached A_ : Optional[Any] = hf_hub_download( _UpperCAmelCase , _UpperCAmelCase , cache_dir=_UpperCAmelCase , force_download=_UpperCAmelCase , proxies=_UpperCAmelCase , resume_download=_UpperCAmelCase , local_files_only=_UpperCAmelCase , use_auth_token=_UpperCAmelCase , ) A_ : str = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment A_ : Optional[int] = check_imports(_UpperCAmelCase ) # Now we move the module inside our cached dynamic modules. A_ : str = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(_UpperCAmelCase ) A_ : Optional[Any] = Path(_UpperCAmelCase ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(_UpperCAmelCase , submodule_path / module_file ) for module_needed in modules_needed: A_ : Tuple = f"""{module_needed}.py""" shutil.copy(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(_UpperCAmelCase , _UpperCAmelCase ): A_ : Union[str, Any] = use_auth_token elif use_auth_token is True: A_ : List[Any] = HfFolder.get_token() else: A_ : Dict = None A_ : List[Any] = model_info(_UpperCAmelCase , revision=_UpperCAmelCase , token=_UpperCAmelCase ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. A_ : Optional[Any] = submodule_path / commit_hash A_ : List[Any] = full_submodule + os.path.sep + commit_hash create_dynamic_module(_UpperCAmelCase ) if not (submodule_path / module_file).exists(): shutil.copy(_UpperCAmelCase , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( _UpperCAmelCase , f"""{module_needed}.py""" , cache_dir=_UpperCAmelCase , force_download=_UpperCAmelCase , resume_download=_UpperCAmelCase , proxies=_UpperCAmelCase , use_auth_token=_UpperCAmelCase , revision=_UpperCAmelCase , local_files_only=_UpperCAmelCase , ) return os.path.join(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = False , **_UpperCAmelCase , ): """simple docstring""" A_ : Optional[Any] = get_cached_module_file( _UpperCAmelCase , _UpperCAmelCase , cache_dir=_UpperCAmelCase , force_download=_UpperCAmelCase , resume_download=_UpperCAmelCase , proxies=_UpperCAmelCase , use_auth_token=_UpperCAmelCase , revision=_UpperCAmelCase , local_files_only=_UpperCAmelCase , ) return get_class_in_module(_UpperCAmelCase , final_module.replace('.py' , '' ) )
286
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ : List[str] = {'configuration_xlnet': ['XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : str = ['XLNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[str] = ['XLNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ 'XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLNetForMultipleChoice', 'XLNetForQuestionAnswering', 'XLNetForQuestionAnsweringSimple', 'XLNetForSequenceClassification', 'XLNetForTokenClassification', 'XLNetLMHeadModel', 'XLNetModel', 'XLNetPreTrainedModel', 'load_tf_weights_in_xlnet', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Union[str, Any] = [ 'TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLNetForMultipleChoice', 'TFXLNetForQuestionAnsweringSimple', 'TFXLNetForSequenceClassification', 'TFXLNetForTokenClassification', 'TFXLNetLMHeadModel', 'TFXLNetMainLayer', 'TFXLNetModel', 'TFXLNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) A_ : Any = hex_num[0] == '-' if is_negative: A_ : Optional[Any] = hex_num[1:] try: A_ : Tuple = int(_UpperCAmelCase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) A_ : Union[str, Any] = '' while int_num > 0: A_ : Optional[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if not is_accelerate_available(): return method A_ : List[Any] = version.parse(accelerate.__version__ ).base_version if version.parse(_UpperCAmelCase ) < version.parse('0.17.0' ): return method def wrapper(self , *_UpperCAmelCase , **_UpperCAmelCase ): if hasattr(self , '_hf_hook' ) and hasattr(self._hf_hook , 'pre_forward' ): self._hf_hook.pre_forward(self ) return method(self , *_UpperCAmelCase , **_UpperCAmelCase ) return wrapper
286
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
1
"""simple docstring""" from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ , snake_case_=1_3 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=9_9 , snake_case_=3_2 , snake_case_=2 , snake_case_=4 , snake_case_=3_7 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=1_6 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): """simple docstring""" A_ : List[str] = parent A_ : Optional[int] = 1_3 A_ : List[str] = 7 A_ : List[Any] = True A_ : Optional[int] = True A_ : Dict = True A_ : List[str] = True A_ : Dict = 9_9 A_ : Optional[Any] = 3_8_4 A_ : str = 2 A_ : Union[str, Any] = 4 A_ : List[str] = 3_7 A_ : str = 'gelu' A_ : Optional[int] = 0.1 A_ : int = 0.1 A_ : List[str] = 5_1_2 A_ : Dict = 1_6 A_ : Union[str, Any] = 2 A_ : Tuple = 0.02 A_ : Any = 3 A_ : List[Any] = 4 A_ : str = 1_2_8 A_ : List[str] = 2 A_ : List[str] = 9 A_ : Union[str, Any] = 1 A_ : Dict = None def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : List[str] = None if self.use_input_mask: A_ : Dict = random_attention_mask([self.batch_size, self.seq_length] ) A_ : Tuple = None if self.use_token_type_ids: A_ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : int = None A_ : Any = None A_ : int = None if self.use_labels: A_ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) A_ : Any = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=snake_case_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = TFConvBertModel(config=snake_case_ ) A_ : str = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} A_ : int = [input_ids, input_mask] A_ : str = model(snake_case_ ) A_ : Any = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = TFConvBertForMaskedLM(config=snake_case_ ) A_ : Tuple = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } A_ : str = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = self.num_labels A_ : int = TFConvBertForSequenceClassification(config=snake_case_ ) A_ : List[str] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } A_ : Any = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = self.num_choices A_ : Union[str, Any] = TFConvBertForMultipleChoice(config=snake_case_ ) A_ : str = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) A_ : str = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) A_ : Optional[int] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) A_ : Union[str, Any] = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } A_ : Any = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : str = self.num_labels A_ : Union[str, Any] = TFConvBertForTokenClassification(config=snake_case_ ) A_ : str = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } A_ : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[str] = TFConvBertForQuestionAnswering(config=snake_case_ ) A_ : Any = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } A_ : List[str] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Tuple = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) : List[Any] = config_and_inputs A_ : List[Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowercase_ : List[Any] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) lowercase_ : Optional[Any] = ( { """feature-extraction""": TFConvBertModel, """fill-mask""": TFConvBertForMaskedLM, """question-answering""": TFConvBertForQuestionAnswering, """text-classification""": TFConvBertForSequenceClassification, """token-classification""": TFConvBertForTokenClassification, """zero-shot""": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) lowercase_ : List[Any] = False lowercase_ : Optional[int] = False lowercase_ : Optional[int] = False def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = TFConvBertModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=snake_case_ , hidden_size=3_7 ) def lowerCamelCase_ ( self ): """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase_ ( self ): """simple docstring""" A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case_ ) @slow def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() A_ : Union[str, Any] = True A_ : str = True if hasattr(snake_case_ , 'use_cache' ): A_ : Optional[int] = True A_ : Optional[Any] = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) A_ : Tuple = getattr(self.model_tester , 'key_length' , snake_case_ ) for model_class in self.all_model_classes: A_ : Union[str, Any] = self._prepare_for_class(snake_case_ , snake_case_ ) A_ : Tuple = model_class(snake_case_ ) A_ : Union[str, Any] = len(model(snake_case_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ , saved_model=snake_case_ ) A_ : Any = os.path.join(snake_case_ , 'saved_model' , '1' ) A_ : Tuple = tf.keras.models.load_model(snake_case_ ) A_ : Optional[int] = model(snake_case_ ) if self.is_encoder_decoder: A_ : str = outputs['encoder_hidden_states'] A_ : Tuple = outputs['encoder_attentions'] else: A_ : int = outputs['hidden_states'] A_ : Any = outputs['attentions'] self.assertEqual(len(snake_case_ ) , snake_case_ ) A_ : List[Any] = getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case_ ) , snake_case_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def lowerCamelCase_ ( self ): """simple docstring""" A_ : Tuple = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) self.assertIsNotNone(snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() A_ : str = True A_ : Optional[int] = getattr(self.model_tester , 'decoder_seq_length' , self.model_tester.seq_length ) A_ : Union[str, Any] = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) A_ : Union[str, Any] = getattr(self.model_tester , 'key_length' , snake_case_ ) A_ : str = getattr(self.model_tester , 'key_length' , snake_case_ ) def check_decoder_attentions_output(snake_case_ ): A_ : str = len(snake_case_ ) self.assertEqual(out_len % 2 , 0 ) A_ : Optional[Any] = outputs.decoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(snake_case_ ): A_ : Tuple = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: A_ : List[str] = True A_ : Any = False A_ : Dict = model_class(snake_case_ ) A_ : int = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) A_ : List[Any] = len(snake_case_ ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) if self.is_encoder_decoder: A_ : Any = model_class(snake_case_ ) A_ : Optional[Any] = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_decoder_attentions_output(snake_case_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] A_ : List[Any] = True A_ : Optional[Any] = model_class(snake_case_ ) A_ : Optional[int] = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) # Check attention is always last and order is fine A_ : Union[str, Any] = True A_ : Dict = True A_ : List[str] = model_class(snake_case_ ) A_ : Any = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(snake_case_ ) ) self.assertEqual(model.config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase_ ( self ): """simple docstring""" A_ : Dict = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) A_ : List[str] = tf.constant([[0, 1, 2, 3, 4, 5]] ) A_ : List[Any] = model(snake_case_ )[0] A_ : int = [1, 6, 7_6_8] self.assertEqual(output.shape , snake_case_ ) A_ : List[Any] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
286
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Any = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """xmod""" def __init__( self , snake_case_=3_0_5_2_2 , snake_case_=7_6_8 , snake_case_=1_2 , snake_case_=1_2 , snake_case_=3_0_7_2 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , snake_case_=False , snake_case_=2 , snake_case_=False , snake_case_=True , snake_case_=True , snake_case_=("en_XX",) , snake_case_=None , **snake_case_ , ): """simple docstring""" super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) A_ : Union[str, Any] = vocab_size A_ : Any = hidden_size A_ : List[str] = num_hidden_layers A_ : Tuple = num_attention_heads A_ : int = hidden_act A_ : Any = intermediate_size A_ : Any = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Union[str, Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : List[str] = initializer_range A_ : Any = layer_norm_eps A_ : Optional[Any] = position_embedding_type A_ : int = use_cache A_ : Dict = classifier_dropout A_ : int = pre_norm A_ : Optional[Any] = adapter_reduction_factor A_ : List[Any] = adapter_layer_norm A_ : int = adapter_reuse_layer_norm A_ : Dict = ln_before_adapter A_ : List[str] = list(snake_case_ ) A_ : Union[str, Any] = default_language class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @property def lowerCamelCase_ ( self ): """simple docstring""" if self.task == "multiple-choice": A_ : Dict = {0: 'batch', 1: 'choice', 2: 'sequence'} else: A_ : int = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
286
1
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging lowerCamelCase_ : Optional[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): """simple docstring""" super().__init__() if hasattr(scheduler.config , 'steps_offset' ) and scheduler.config.steps_offset != 1: A_ : Dict = ( F"""The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`""" F""" should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure """ 'to update the config accordingly as leaving `steps_offset` might led to incorrect results' ' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,' ' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`' ' file' ) deprecate('steps_offset!=1' , '1.0.0' , snake_case_ , standard_warn=snake_case_ ) A_ : Optional[Any] = dict(scheduler.config ) A_ : List[Any] = 1 A_ : int = FrozenDict(snake_case_ ) if hasattr(scheduler.config , 'skip_prk_steps' ) and scheduler.config.skip_prk_steps is False: A_ : Dict = ( F"""The configuration file of this scheduler: {scheduler} has not set the configuration""" ' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make' ' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to' ' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face' ' Hub, it would be very nice if you could open a Pull request for the' ' `scheduler/scheduler_config.json` file' ) deprecate('skip_prk_steps not set' , '1.0.0' , snake_case_ , standard_warn=snake_case_ ) A_ : List[Any] = dict(scheduler.config ) A_ : Optional[Any] = True A_ : List[Any] = FrozenDict(snake_case_ ) if safety_checker is None: logger.warning( F"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( segmentation_model=snake_case_ , segmentation_processor=snake_case_ , vae=snake_case_ , text_encoder=snake_case_ , tokenizer=snake_case_ , unet=snake_case_ , scheduler=snake_case_ , safety_checker=snake_case_ , feature_extractor=snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ = "auto" ): """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory A_ : Tuple = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" self.enable_attention_slicing(snake_case_ ) def lowerCamelCase_ ( self ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) A_ : Optional[int] = torch.device('cuda' ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(snake_case_ , snake_case_ ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCamelCase_ ( self ): """simple docstring""" if self.device != torch.device('meta' ) or not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(snake_case_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ = 5_1_2 , snake_case_ = 5_1_2 , snake_case_ = 5_0 , snake_case_ = 7.5 , snake_case_ = None , snake_case_ = 1 , snake_case_ = 0.0 , snake_case_ = None , snake_case_ = None , snake_case_ = "pil" , snake_case_ = True , snake_case_ = None , snake_case_ = 1 , **snake_case_ , ): """simple docstring""" A_ : List[str] = self.segmentation_processor( text=[text] , images=[image] , padding='max_length' , return_tensors='pt' ).to(self.device ) A_ : Optional[Any] = self.segmentation_model(**snake_case_ ) A_ : Tuple = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() A_ : List[Any] = self.numpy_to_pil(snake_case_ )[0].resize(image.size ) # Run inpainting pipeline with the generated mask A_ : List[Any] = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=snake_case_ , image=snake_case_ , mask_image=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , )
286
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
1
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if days_between_payments <= 0: raise ValueError('days_between_payments must be > 0' ) if daily_interest_rate < 0: raise ValueError('daily_interest_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * daily_interest_rate * days_between_payments def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" if number_of_compounding_periods <= 0: raise ValueError('number_of_compounding_periods must be > 0' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('nominal_annual_interest_rate_percentage must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" if number_of_years <= 0: raise ValueError('number_of_years must be > 0' ) if nominal_annual_percentage_rate < 0: raise ValueError('nominal_annual_percentage_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return compound_interest( _UpperCAmelCase , nominal_annual_percentage_rate / 365 , number_of_years * 365 ) if __name__ == "__main__": import doctest doctest.testmod()
286
"""simple docstring""" from copy import deepcopy class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None , snake_case_ = None ): """simple docstring""" if arr is None and size is not None: A_ : Union[str, Any] = size A_ : List[str] = [0] * size elif arr is not None: self.init(snake_case_ ) else: raise ValueError('Either arr or size must be specified' ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = len(snake_case_ ) A_ : Optional[int] = deepcopy(snake_case_ ) for i in range(1 , self.size ): A_ : Optional[Any] = self.next_(snake_case_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): A_ : Optional[int] = self.next_(snake_case_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index + (index & (-index)) @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index - (index & (-index)) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value A_ : List[str] = self.next_(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" self.add(snake_case_ , value - self.get(snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if right == 0: return 0 A_ : Any = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] A_ : Tuple = self.prev(snake_case_ ) return result def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" return self.prefix(snake_case_ ) - self.prefix(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.query(snake_case_ , index + 1 ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" value -= self.tree[0] if value < 0: return -1 A_ : List[Any] = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 A_ : Tuple = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" from __future__ import annotations from math import gcd def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase = 2 , _UpperCAmelCase = 1 , _UpperCAmelCase = 3 , ): """simple docstring""" if num < 2: raise ValueError('The input value cannot be less than 2' ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: return (pow(_UpperCAmelCase , 2 ) + step) % modulus for _ in range(_UpperCAmelCase ): # These track the position within the cycle detection logic. A_ : List[Any] = seed A_ : Any = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. A_ : Union[str, Any] = rand_fn(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : Tuple = rand_fn(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : List[str] = rand_fn(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. A_ : Any = gcd(hare - tortoise , _UpperCAmelCase ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. A_ : Any = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse lowerCamelCase_ : int = argparse.ArgumentParser() parser.add_argument( 'num', type=int, help='The value to find a divisor of', ) parser.add_argument( '--attempts', type=int, default=3, help='The number of attempts before giving up', ) lowerCamelCase_ : Optional[int] = parser.parse_args() lowerCamelCase_ : List[str] = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(F"{args.num} is probably prime") else: lowerCamelCase_ : List[str] = args.num // divisor print(F"{args.num} = {divisor} * {quotient}")
286
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , snake_case_ = 7_6_8 , ): """simple docstring""" super().__init__() A_ : Optional[int] = nn.Parameter(torch.zeros(1 , snake_case_ ) ) A_ : Optional[int] = nn.Parameter(torch.ones(1 , snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ = None , snake_case_ = None , ): """simple docstring""" A_ : str = nn.Parameter(self.mean.to(snake_case_ ).to(snake_case_ ) ) A_ : Optional[int] = nn.Parameter(self.std.to(snake_case_ ).to(snake_case_ ) ) return self def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Tuple = (embeds - self.mean) * 1.0 / self.std return embeds def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = (embeds * self.std) + self.mean return embeds
286
1
"""simple docstring""" # Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=0 ): """simple docstring""" if name is None: A_ : Optional[Any] = None else: A_ : Dict = '.' * max(0 , spaces - 2 ) + '# {:' + str(50 - spaces ) + 's}' A_ : List[str] = fmt.format(_UpperCAmelCase ) # Print and recurse (if needed). if isinstance(_UpperCAmelCase , _UpperCAmelCase ): if msg is not None: print(_UpperCAmelCase ) for k in val.keys(): recursive_print(_UpperCAmelCase , val[k] , spaces + 2 ) elif isinstance(_UpperCAmelCase , torch.Tensor ): print(_UpperCAmelCase , ':' , val.size() ) else: print(_UpperCAmelCase , ':' , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] A_ : Optional[Any] = (num_heads, hidden_size, num_splits) + input_shape[1:] A_ : Tuple = param.view(*_UpperCAmelCase ) A_ : List[Any] = param.transpose(0 , 2 ) A_ : List[Any] = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] A_ : Any = (num_heads, num_splits, hidden_size) + input_shape[1:] A_ : List[str] = param.view(*_UpperCAmelCase ) A_ : Union[str, Any] = param.transpose(0 , 1 ).contiguous() A_ : Any = param.view(*_UpperCAmelCase ) return param def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : int = {} # old versions did not store training args A_ : Optional[Any] = input_state_dict.get('args' , _UpperCAmelCase ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) A_ : str = ds_args.padded_vocab_size A_ : Tuple = ds_args.max_position_embeddings A_ : int = ds_args.hidden_size A_ : str = ds_args.num_layers A_ : List[Any] = ds_args.num_attention_heads A_ : str = ds_args.ffn_hidden_size # pprint(config) # The number of heads. A_ : str = config.n_head # The hidden_size per head. A_ : Union[str, Any] = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): A_ : Dict = input_state_dict['checkpoint_version'] else: A_ : Optional[Any] = 0.0 # The model. A_ : Optional[Any] = input_state_dict['model'] # The language model. A_ : List[str] = model['language_model'] # The embeddings. A_ : Optional[Any] = lm['embedding'] # The word embeddings. A_ : Tuple = embeddings['word_embeddings']['weight'] # Truncate the embedding table to vocab_size rows. A_ : Union[str, Any] = word_embeddings[: config.vocab_size, :] A_ : Union[str, Any] = word_embeddings # The position embeddings. A_ : Tuple = embeddings['position_embeddings']['weight'] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] A_ : Tuple = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( f"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don't match""" ) # Store the position embeddings. A_ : Dict = pos_embeddings # The transformer. A_ : Optional[int] = lm['transformer'] if 'transformer' in lm.keys() else lm['encoder'] # The regex to extract layer names. A_ : List[Any] = re.compile(R'layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)' ) # The simple map of names for "automated" rules. A_ : Any = { 'attention.dense': '.attn.c_proj.', 'self_attention.dense': '.attn.c_proj.', 'mlp.dense_h_to_4h': '.mlp.c_fc.', 'mlp.dense_4h_to_h': '.mlp.c_proj.', } # Extract the layers. for key, val in transformer.items(): # Match the name. A_ : List[Any] = layer_re.match(_UpperCAmelCase ) # Stop if that's not a layer if m is None: break # The index of the layer. A_ : Union[str, Any] = int(m.group(1 ) ) # The name of the operation. A_ : Union[str, Any] = m.group(2 ) # Is it a weight or a bias? A_ : int = m.group(3 ) # The name of the layer. A_ : Union[str, Any] = f"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith('layernorm' ): A_ : Any = 'ln_1' if op_name.startswith('input' ) else 'ln_2' A_ : Union[str, Any] = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. A_ : str = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , _UpperCAmelCase , _UpperCAmelCase ) A_ : Union[str, Any] = causal_mask # Insert a "dummy" tensor for masked_bias. A_ : Optional[Any] = torch.tensor(-1E4 , dtype=torch.floataa ) A_ : Any = masked_bias A_ : List[str] = fix_query_key_value_ordering(_UpperCAmelCase , _UpperCAmelCase , 3 , _UpperCAmelCase , _UpperCAmelCase ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. A_ : Any = out_val.transpose(0 , 1 ).contiguous() # Store. A_ : Any = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": A_ : Dict = fix_query_key_value_ordering(_UpperCAmelCase , _UpperCAmelCase , 3 , _UpperCAmelCase , _UpperCAmelCase ) # Store. No change of shape. A_ : int = out_val # Transpose the weights. elif weight_or_bias == "weight": A_ : Union[str, Any] = megatron_to_transformers[op_name] A_ : Optional[Any] = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": A_ : Optional[Any] = megatron_to_transformers[op_name] A_ : Any = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. A_ : List[Any] = transformer['final_layernorm.weight'] A_ : Union[str, Any] = transformer['final_layernorm.bias'] # For LM head, transformers' wants the matrix to weight embeddings. A_ : List[str] = word_embeddings # It should be done! return output_state_dict def UpperCAmelCase__ ( ): """simple docstring""" A_ : Any = argparse.ArgumentParser() parser.add_argument('--print-checkpoint-structure' , action='store_true' ) parser.add_argument( 'path_to_checkpoint' , type=_UpperCAmelCase , help='Path to the checkpoint file (.zip archive or direct .pt file)' , ) parser.add_argument( '--config_file' , default='' , type=_UpperCAmelCase , help='An optional config json file describing the pre-trained model.' , ) A_ : Optional[Any] = parser.parse_args() # Extract the basename. A_ : Any = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(f"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""" ) if args.path_to_checkpoint.endswith('.zip' ): with zipfile.ZipFile(args.path_to_checkpoint , 'r' ) as checkpoint: with checkpoint.open('release/mp_rank_00/model_optim_rng.pt' ) as pytorch_dict: A_ : List[str] = torch.load(_UpperCAmelCase , map_location='cpu' ) else: A_ : int = torch.load(args.path_to_checkpoint , map_location='cpu' ) A_ : Dict = input_state_dict.get('args' , _UpperCAmelCase ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: A_ : int = 'gelu_fast' elif ds_args.openai_gelu: A_ : Any = 'gelu_new' else: A_ : Dict = 'gelu' else: # in the very early days this used to be "gelu_new" A_ : Dict = 'gelu_new' # Spell out all parameters in case the defaults change. A_ : Any = GPTaConfig( vocab_size=50257 , n_positions=1024 , n_embd=1024 , n_layer=24 , n_head=16 , n_inner=4096 , activation_function=_UpperCAmelCase , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1E-5 , initializer_range=0.02 , summary_type='cls_index' , summary_use_proj=_UpperCAmelCase , summary_activation=_UpperCAmelCase , summary_proj_to_labels=_UpperCAmelCase , summary_first_dropout=0.1 , scale_attn_weights=_UpperCAmelCase , use_cache=_UpperCAmelCase , bos_token_id=50256 , eos_token_id=50256 , ) else: A_ : List[str] = GPTaConfig.from_json_file(args.config_file ) A_ : Optional[Any] = ['GPT2LMHeadModel'] # Convert. print('Converting' ) A_ : Union[str, Any] = convert_megatron_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(_UpperCAmelCase , _UpperCAmelCase ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: A_ : List[Any] = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": A_ : Optional[Any] = 'gpt2' elif tokenizer_type == "PretrainedFromHF": A_ : str = ds_args.tokenizer_name_or_path else: raise ValueError(f"""Unrecognized tokenizer_type {tokenizer_type}""" ) else: A_ : Tuple = 'gpt2' A_ : Any = AutoTokenizer.from_pretrained(_UpperCAmelCase ) A_ : Dict = type(_UpperCAmelCase ).__name__ A_ : int = tokenizer_class # Store the config to file. print('Saving config' ) config.save_pretrained(_UpperCAmelCase ) # Save tokenizer based on args print(f"""Adding {tokenizer_class} tokenizer files""" ) tokenizer.save_pretrained(_UpperCAmelCase ) # Store the state_dict to file. A_ : List[Any] = os.path.join(_UpperCAmelCase , 'pytorch_model.bin' ) print(f"""Saving checkpoint to \"{output_checkpoint_file}\"""" ) torch.save(_UpperCAmelCase , _UpperCAmelCase ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
286
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCamelCase_ : Any = HfArgumentParser(InitializationArguments) lowerCamelCase_ : Union[str, Any] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowerCamelCase_ : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowerCamelCase_ : Tuple = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) lowerCamelCase_ : int = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowerCamelCase_ : Any = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
286
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer lowerCamelCase_ : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase_ : Tuple = '\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu")\n\n >>> repo = "openai/shap-e-img2img"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png"\n >>> image = load_image(image_url).convert("RGB")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], "corgi_3d.gif")\n ```\n' @dataclass class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Union[PIL.Image.Image, np.ndarray] class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): """simple docstring""" super().__init__() self.register_modules( prior=snake_case_ , image_encoder=snake_case_ , image_processor=snake_case_ , scheduler=snake_case_ , renderer=snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" if latents is None: A_ : List[Any] = randn_tensor(snake_case_ , generator=snake_case_ , device=snake_case_ , dtype=snake_case_ ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) A_ : Any = latents.to(snake_case_ ) A_ : str = latents * scheduler.init_noise_sigma return latents def lowerCamelCase_ ( self , snake_case_=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) A_ : Any = torch.device(F"""cuda:{gpu_id}""" ) A_ : Optional[Any] = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(snake_case_ , snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" if self.device != torch.device('meta' ) or not hasattr(self.image_encoder , '_hf_hook' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(snake_case_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def lowerCamelCase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ) and isinstance(image[0] , torch.Tensor ): A_ : List[Any] = torch.cat(snake_case_ , axis=0 ) if image[0].ndim == 4 else torch.stack(snake_case_ , axis=0 ) if not isinstance(snake_case_ , torch.Tensor ): A_ : str = self.image_processor(snake_case_ , return_tensors='pt' ).pixel_values[0].unsqueeze(0 ) A_ : Union[str, Any] = image.to(dtype=self.image_encoder.dtype , device=snake_case_ ) A_ : Any = self.image_encoder(snake_case_ )['last_hidden_state'] A_ : List[Any] = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 A_ : List[str] = image_embeds.repeat_interleave(snake_case_ , dim=0 ) if do_classifier_free_guidance: A_ : Tuple = torch.zeros_like(snake_case_ ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes A_ : str = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(snake_case_ ) def __call__( self , snake_case_ , snake_case_ = 1 , snake_case_ = 2_5 , snake_case_ = None , snake_case_ = None , snake_case_ = 4.0 , snake_case_ = 6_4 , snake_case_ = "pil" , snake_case_ = True , ): """simple docstring""" if isinstance(snake_case_ , PIL.Image.Image ): A_ : Any = 1 elif isinstance(snake_case_ , torch.Tensor ): A_ : Any = image.shape[0] elif isinstance(snake_case_ , snake_case_ ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): A_ : Union[str, Any] = len(snake_case_ ) else: raise ValueError( F"""`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(snake_case_ )}""" ) A_ : Any = self._execution_device A_ : Optional[int] = batch_size * num_images_per_prompt A_ : int = guidance_scale > 1.0 A_ : Optional[int] = self._encode_image(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # prior self.scheduler.set_timesteps(snake_case_ , device=snake_case_ ) A_ : int = self.scheduler.timesteps A_ : Tuple = self.prior.config.num_embeddings A_ : List[str] = self.prior.config.embedding_dim A_ : int = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , snake_case_ , snake_case_ , snake_case_ , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim A_ : List[str] = latents.reshape(latents.shape[0] , snake_case_ , snake_case_ ) for i, t in enumerate(self.progress_bar(snake_case_ ) ): # expand the latents if we are doing classifier free guidance A_ : str = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents A_ : List[Any] = self.scheduler.scale_model_input(snake_case_ , snake_case_ ) A_ : Union[str, Any] = self.prior( snake_case_ , timestep=snake_case_ , proj_embedding=snake_case_ , ).predicted_image_embedding # remove the variance A_ , A_ : Tuple = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: A_ , A_ : List[Any] = noise_pred.chunk(2 ) A_ : List[Any] = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) A_ : List[Any] = self.scheduler.step( snake_case_ , timestep=snake_case_ , sample=snake_case_ , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=snake_case_ ) A_ : Optional[Any] = [] for i, latent in enumerate(snake_case_ ): print() A_ : str = self.renderer.decode( latent[None, :] , snake_case_ , size=snake_case_ , ray_batch_size=4_0_9_6 , n_coarse_samples=6_4 , n_fine_samples=1_2_8 , ) images.append(snake_case_ ) A_ : Union[str, Any] = torch.stack(snake_case_ ) if output_type not in ["np", "pil"]: raise ValueError(F"""Only the output types `pil` and `np` are supported not output_type={output_type}""" ) A_ : str = images.cpu().numpy() if output_type == "pil": A_ : Optional[Any] = [self.numpy_to_pil(snake_case_ ) for image in images] # Offload last model to CPU if hasattr(self , 'final_offload_hook' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=snake_case_ )
286
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser lowerCamelCase_ : Any = re.compile(r'\s+') def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return {"hash": hashlib.mda(re.sub(_UpperCAmelCase , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [len(_UpperCAmelCase ) for line in example['content'].splitlines()] return {"line_mean": np.mean(_UpperCAmelCase ), "line_max": max(_UpperCAmelCase )} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 ): """simple docstring""" A_ : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] A_ : List[str] = example['content'].splitlines() for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 , _UpperCAmelCase=0.05 ): """simple docstring""" A_ : Any = ['unit tests', 'test file', 'configuration file'] A_ : Dict = example['content'].splitlines() A_ : List[Any] = 0 A_ : str = 0 # first test for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test A_ : Tuple = example['content'].count('\n' ) A_ : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = ['def ', 'class ', 'for ', 'while '] A_ : Tuple = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=4 ): """simple docstring""" A_ : Union[str, Any] = example['content'].splitlines() A_ : Any = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = tokenizer(example['content'] , truncation=_UpperCAmelCase )['input_ids'] A_ : Dict = len(example['content'] ) / len(_UpperCAmelCase ) return {"ratio": ratio} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = {} results.update(get_hash(_UpperCAmelCase ) ) results.update(line_stats(_UpperCAmelCase ) ) results.update(alpha_stats(_UpperCAmelCase ) ) results.update(char_token_ratio(_UpperCAmelCase ) ) results.update(is_autogenerated(_UpperCAmelCase ) ) results.update(is_config_or_test(_UpperCAmelCase ) ) results.update(has_no_keywords(_UpperCAmelCase ) ) results.update(has_few_assignments(_UpperCAmelCase ) ) return results def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if not check_uniques(_UpperCAmelCase , _UpperCAmelCase ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'rb' ) as f_in: with gzip.open(str(_UpperCAmelCase ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) os.unlink(_UpperCAmelCase ) # Settings lowerCamelCase_ : Optional[int] = HfArgumentParser(PreprocessingArguments) lowerCamelCase_ : Optional[Any] = parser.parse_args() if args.num_workers is None: lowerCamelCase_ : int = multiprocessing.cpu_count() lowerCamelCase_ : Tuple = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset lowerCamelCase_ : Tuple = time.time() lowerCamelCase_ : Tuple = load_dataset(args.dataset_name, split='train') print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing lowerCamelCase_ : List[str] = time.time() lowerCamelCase_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes lowerCamelCase_ : int = set(ds.unique('hash')) lowerCamelCase_ : Union[str, Any] = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics lowerCamelCase_ : Optional[int] = time.time() lowerCamelCase_ : Tuple = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: lowerCamelCase_ : Union[str, Any] = time.time() lowerCamelCase_ , lowerCamelCase_ : str = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file lowerCamelCase_ : Tuple = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) lowerCamelCase_ : Optional[Any] = output_dir / 'data' data_dir.mkdir(exist_ok=True) lowerCamelCase_ : List[str] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): lowerCamelCase_ : Optional[int] = str(data_dir / F"file-{file_number+1:012}.json") lowerCamelCase_ : List[str] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
286
1
"""simple docstring""" from collections.abc import Iterable from typing import Any class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None ): """simple docstring""" A_ : Optional[Any] = value A_ : Node | None = None # Added in order to delete a node easier A_ : Node | None = None A_ : Node | None = None def __repr__( self ): """simple docstring""" from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F"""{self.value}""": (self.left, self.right)} , indent=1 ) class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None ): """simple docstring""" A_ : int = root def __str__( self ): """simple docstring""" return str(self.root ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if new_children is not None: # reset its kids A_ : Any = node.parent if node.parent is not None: # reset its parent if self.is_right(snake_case_ ): # If it is the right children A_ : int = new_children else: A_ : Union[str, Any] = new_children else: A_ : Optional[Any] = new_children def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if node.parent and node.parent.right: return node == node.parent.right return False def lowerCamelCase_ ( self ): """simple docstring""" return self.root is None def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : int = Node(snake_case_ ) # create a new Node if self.empty(): # if Tree is empty A_ : Any = new_node # set its root else: # Tree is not empty A_ : List[str] = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: A_ : Optional[Any] = new_node # We insert the new node in a leaf break else: A_ : Union[str, Any] = parent_node.left else: if parent_node.right is None: A_ : Any = new_node break else: A_ : Tuple = parent_node.right A_ : Tuple = parent_node def lowerCamelCase_ ( self , *snake_case_ ): """simple docstring""" for value in values: self.__insert(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.empty(): raise IndexError('Warning: Tree is empty! please use another.' ) else: A_ : Any = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: A_ : str = node.left if value < node.value else node.right return node def lowerCamelCase_ ( self , snake_case_ = None ): """simple docstring""" if node is None: if self.root is None: return None A_ : str = self.root if not self.empty(): while node.right is not None: A_ : List[str] = node.right return node def lowerCamelCase_ ( self , snake_case_ = None ): """simple docstring""" if node is None: A_ : str = self.root if self.root is None: return None if not self.empty(): A_ : Optional[Any] = self.root while node.left is not None: A_ : Dict = node.left return node def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Optional[Any] = self.search(snake_case_ ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(snake_case_ , snake_case_ ) elif node.left is None: # Has only right children self.__reassign_nodes(snake_case_ , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(snake_case_ , node.left ) else: A_ : Tuple = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore A_ : Optional[Any] = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def lowerCamelCase_ ( self , snake_case_=None ): """simple docstring""" if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if node: self.inorder(snake_case_ , node.left ) arr.append(node.value ) self.inorder(snake_case_ , node.right ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : list[int] = [] self.inorder(snake_case_ , snake_case_ ) # append all values to list using inorder traversal return arr[k - 1] def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[int] = [] if curr_node is not None: A_ : Tuple = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def UpperCAmelCase__ ( ): """simple docstring""" A_ : int = (8, 3, 6, 1, 10, 14, 13, 4, 7) A_ : Optional[int] = BinarySearchTree() for i in testlist: t.insert(_UpperCAmelCase ) # Prints all the elements of the list in order traversal print(_UpperCAmelCase ) if t.search(6 ) is not None: print('The value 6 exists' ) else: print('The value 6 doesn\'t exist' ) if t.search(-1 ) is not None: print('The value -1 exists' ) else: print('The value -1 doesn\'t exist' ) if not t.empty(): print('Max Value: ' , t.get_max().value ) # type: ignore print('Min Value: ' , t.get_min().value ) # type: ignore for i in testlist: t.remove(_UpperCAmelCase ) print(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
286
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : Optional[Any] = [] # fmt: off # stem: rename_keys.append(('cls_token', 'vit.embeddings.cls_token') ) rename_keys.append(('pos_embed', 'vit.embeddings.position_embeddings') ) rename_keys.append(('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias') ) # backbone rename_keys.append(('patch_embed.backbone.stem.conv.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.bias', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A_ : List[str] = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) # fmt: on return rename_keys def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: A_ : List[str] = '' else: A_ : Dict = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ : str = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) A_ : List[Any] = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A_ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A_ : Tuple = in_proj_bias[: config.hidden_size] A_ : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A_ : Tuple = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = dct.pop(_UpperCAmelCase ) A_ : Optional[int] = val def UpperCAmelCase__ ( ): """simple docstring""" A_ : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' A_ : int = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : List[Any] = BitConfig( global_padding='same' , layer_type='bottleneck' , depths=(3, 4, 9) , out_features=['stage3'] , embedding_dynamic_padding=_UpperCAmelCase , ) A_ : Optional[int] = ViTHybridConfig(backbone_config=_UpperCAmelCase , image_size=384 , num_labels=1000 ) A_ : Union[str, Any] = False # load original model from timm A_ : List[Any] = timm.create_model(_UpperCAmelCase , pretrained=_UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ : Tuple = timm_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) A_ : Any = create_rename_keys(_UpperCAmelCase , _UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : Union[str, Any] = 'huggingface/label-files' A_ : Dict = 'imagenet-1k-id2label.json' A_ : List[str] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) A_ : str = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} A_ : Any = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": A_ : List[Any] = ViTHybridModel(_UpperCAmelCase ).eval() else: A_ : str = ViTHybridForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # create image processor A_ : Dict = create_transform(**resolve_data_config({} , model=_UpperCAmelCase ) ) A_ : List[str] = transform.transforms A_ : List[str] = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } A_ : Tuple = ViTHybridImageProcessor( do_resize=_UpperCAmelCase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_UpperCAmelCase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=_UpperCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) A_ : Optional[Any] = prepare_img() A_ : Any = transform(_UpperCAmelCase ).unsqueeze(0 ) A_ : Dict = processor(_UpperCAmelCase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase ) # verify logits with torch.no_grad(): A_ : List[Any] = model(_UpperCAmelCase ) A_ : List[str] = outputs.logits print('Predicted class:' , logits.argmax(-1 ).item() ) if base_model: A_ : Union[str, Any] = timm_model.forward_features(_UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_UpperCAmelCase , outputs.pooler_output , atol=1E-3 ) else: A_ : Tuple = timm_model(_UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase , outputs.logits , atol=1E-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(f"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: print(f"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(f"""ybelkada/{vit_name}""" ) processor.push_to_hub(f"""ybelkada/{vit_name}""" ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_r50_s16_384', type=str, help='Name of the hybrid ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) lowerCamelCase_ : List[str] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
286
1
"""simple docstring""" def UpperCAmelCase__ ( ): """simple docstring""" return 1 def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else five_pence(x - 5 ) + two_pence(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return 0 if x < 0 else two_pound(x - 200 ) + one_pound(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase = 200 ): """simple docstring""" return two_pound(_UpperCAmelCase ) if __name__ == "__main__": print(solution(int(input().strip())))
286
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" A_ : str = False if num < 0: A_ : Dict = True A_ : Union[str, Any] = -num A_ : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(_UpperCAmelCase ) for e in binary ) return "0b" + "".join(str(_UpperCAmelCase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
286
1
"""simple docstring""" lowerCamelCase_ : Union[str, Any] = range(2, 20 + 1) lowerCamelCase_ : Dict = [10**k for k in range(ks[-1] + 1)] lowerCamelCase_ : dict[int, dict[int, list[list[int]]]] = {} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[str] = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) A_ : Any = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) A_ , A_ : List[Any] = 0, 0 A_ : List[Any] = n - i A_ : Optional[Any] = memo.get(_UpperCAmelCase ) if sub_memo is not None: A_ : Union[str, Any] = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over A_ : Dict = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: A_ : str = _k break if max_jump >= 0: A_ , A_ , A_ : Tuple = jumps[max_jump] # since the difference between jumps is cached, add c A_ : Any = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): A_ , A_ : int = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: A_ : str = [] else: A_ : List[str] = {c: []} A_ : Union[str, Any] = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps A_ , A_ : Optional[int] = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead A_ , A_ : Optional[int] = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped A_ : Optional[int] = sub_memo[c] # keep jumps sorted by # of terms skipped A_ : Optional[int] = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) A_ : List[Any] = i A_ , A_ , A_ : List[str] = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 A_ : str = ds_c + ds_b diff += addend A_ : int = 0 for j in range(_UpperCAmelCase ): A_ : Dict = a_i[j] + addend A_ , A_ : List[Any] = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): A_ : Dict = digits[j] + addend if s >= 10: A_ , A_ : Union[str, Any] = divmod(_UpperCAmelCase , 10 ) A_ : Dict = addend // 10 + quotient else: A_ : int = s A_ : Any = addend // 10 if addend == 0: break while addend > 0: A_ , A_ : Any = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase = 10**15 ): """simple docstring""" A_ : str = [1] A_ : Any = 1 A_ : Dict = 0 while True: A_ , A_ : Optional[int] = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break A_ : Any = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F"{solution() = }")
286
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCamelCase_ : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Tuple = ['MLukeTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys lowerCamelCase_ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
1
"""simple docstring""" import os def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : int = len(grid[0] ) A_ : str = len(_UpperCAmelCase ) A_ : int = 0 A_ : int = 0 A_ : Union[str, Any] = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(_UpperCAmelCase ): for j in range(n_rows - 3 ): A_ : Tuple = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] A_ : List[str] = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: A_ : Any = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: A_ : Tuple = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) A_ : List[Any] = max( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if max_product > largest: A_ : List[str] = max_product return largest def UpperCAmelCase__ ( ): """simple docstring""" A_ : Optional[int] = [] with open(os.path.dirname(_UpperCAmelCase ) + '/grid.txt' ) as file: for line in file: grid.append(line.strip('\n' ).split(' ' ) ) A_ : Tuple = [[int(_UpperCAmelCase ) for i in grid[j]] for j in range(len(_UpperCAmelCase ) )] return largest_product(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
286
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers lowerCamelCase_ : List[str] = [int(0.5 * n * (n + 1)) for n in range(1, 1_01)] def UpperCAmelCase__ ( ): """simple docstring""" A_ : Union[str, Any] = os.path.dirname(os.path.realpath(_UpperCAmelCase ) ) A_ : Tuple = os.path.join(_UpperCAmelCase , 'words.txt' ) A_ : List[Any] = '' with open(_UpperCAmelCase ) as f: A_ : int = f.readline() A_ : Optional[Any] = [word.strip('"' ) for word in words.strip('\r\n' ).split(',' )] A_ : Dict = [ word for word in [sum(ord(_UpperCAmelCase ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
286
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class lowercase_ ( unittest.TestCase ): '''simple docstring''' __snake_case = ViTImageProcessor if is_vision_available() else None @property def __lowerCAmelCase ( self : Dict ) ->Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self : Dict ) ->int: """simple docstring""" a = (3, 32, 128) a = tempfile.mkdtemp() # fmt: off a = ['''[GO]''', '''[s]''', '''0''', '''1''', '''2''', '''3''', '''4''', '''5''', '''6''', '''7''', '''8''', '''9''', '''a''', '''b''', '''c''', '''d''', '''e''', '''f''', '''g''', '''h''', '''i''', '''j''', '''k''', '''l''', '''m''', '''n''', '''o''', '''p''', '''q''', '''r''', '''s''', '''t''', '''u''', '''v''', '''w''', '''x''', '''y''', '''z'''] # fmt: on a = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + '''\n''' ) a = { '''do_normalize''': False, '''do_resize''': True, '''image_processor_type''': '''ViTImageProcessor''', '''resample''': 3, '''size''': {'''height''': 32, '''width''': 128}, } a = os.path.join(self.tmpdirname , __UpperCAmelCase ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(__UpperCAmelCase , __UpperCAmelCase ) def __lowerCAmelCase ( self : Union[str, Any] , **__UpperCAmelCase : List[Any] ) ->List[Any]: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__UpperCAmelCase ) def __lowerCAmelCase ( self : List[Any] , **__UpperCAmelCase : Tuple ) ->List[str]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ) def __lowerCAmelCase ( self : Optional[int] ) ->List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : int ) ->Dict: """simple docstring""" a = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta ) a = Image.fromarray(np.moveaxis(__UpperCAmelCase , 0 , -1 ) ) return image_input def __lowerCAmelCase ( self : Optional[Any] ) ->str: """simple docstring""" a = self.get_tokenizer() a = self.get_image_processor() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) a = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__UpperCAmelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __UpperCAmelCase ) def __lowerCAmelCase ( self : Tuple ) ->str: """simple docstring""" a = self.get_tokenizer() a = self.get_image_processor() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) a = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) a = self.get_image_processor(do_normalize=__UpperCAmelCase , padding_value=1.0 ) a = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __UpperCAmelCase ) def __lowerCAmelCase ( self : Tuple ) ->str: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) a = self.prepare_image_inputs() a = image_processor(__UpperCAmelCase , return_tensors='''np''' ) a = processor(images=__UpperCAmelCase , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __lowerCAmelCase ( self : List[str] ) ->Any: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) a = '''test''' a = processor(text=__UpperCAmelCase ) a = tokenizer(__UpperCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __lowerCAmelCase ( self : Union[str, Any] ) ->Tuple: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) a = '''test''' a = self.prepare_image_inputs() a = processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ['''pixel_values''', '''labels'''] ) # test if it raises when no input is passed with pytest.raises(__UpperCAmelCase ): processor() def __lowerCAmelCase ( self : str ) ->Dict: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) a = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] a = processor.char_decode(__UpperCAmelCase ) a = tokenizer.batch_decode(__UpperCAmelCase ) a = [seq.replace(''' ''' , '''''' ) for seq in decoded_tok] self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def __lowerCAmelCase ( self : int ) ->int: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) a = None a = self.prepare_image_inputs() a = processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def __lowerCAmelCase ( self : Any ) ->Any: """simple docstring""" a = self.get_image_processor() a = self.get_tokenizer() a = MgpstrProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) a = torch.randn(1 , 27 , 38 ) a = torch.randn(1 , 27 , 50_257 ) a = torch.randn(1 , 27 , 30_522 ) a = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['''generated_text''', '''scores''', '''char_preds''', '''bpe_preds''', '''wp_preds'''] )
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ : List[str] = {'configuration_xlnet': ['XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : str = ['XLNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[str] = ['XLNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ 'XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLNetForMultipleChoice', 'XLNetForQuestionAnswering', 'XLNetForQuestionAnsweringSimple', 'XLNetForSequenceClassification', 'XLNetForTokenClassification', 'XLNetLMHeadModel', 'XLNetModel', 'XLNetPreTrainedModel', 'load_tf_weights_in_xlnet', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Union[str, Any] = [ 'TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLNetForMultipleChoice', 'TFXLNetForQuestionAnsweringSimple', 'TFXLNetForSequenceClassification', 'TFXLNetForTokenClassification', 'TFXLNetLMHeadModel', 'TFXLNetMainLayer', 'TFXLNetModel', 'TFXLNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
0
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def lowerCAmelCase_ ( snake_case_ : int ) -> List[str]: '''simple docstring''' if "img_encoder.pos_embed" in name: UpperCAmelCase_ = name.replace("img_encoder.pos_embed" , "vision_model.embeddings.position_embeddings" ) if "img_encoder.patch_embed.proj" in name: UpperCAmelCase_ = name.replace("img_encoder.patch_embed.proj" , "vision_model.embeddings.patch_embeddings.projection" ) if "img_encoder.patch_embed.norm" in name: UpperCAmelCase_ = name.replace("img_encoder.patch_embed.norm" , "vision_model.embeddings.layernorm" ) if "img_encoder.layers" in name: UpperCAmelCase_ = name.replace("img_encoder.layers" , "vision_model.encoder.stages" ) if "blocks" in name and "res" not in name: UpperCAmelCase_ = name.replace("blocks" , "layers" ) if "attn" in name and "pre_assign" not in name: UpperCAmelCase_ = name.replace("attn" , "self_attn" ) if "proj" in name and "self_attn" in name and "text" not in name: UpperCAmelCase_ = name.replace("proj" , "out_proj" ) if "pre_assign_attn.attn.proj" in name: UpperCAmelCase_ = name.replace("pre_assign_attn.attn.proj" , "pre_assign_attn.attn.out_proj" ) if "norm1" in name: UpperCAmelCase_ = name.replace("norm1" , "layer_norm1" ) if "norm2" in name and "pre_assign" not in name: UpperCAmelCase_ = name.replace("norm2" , "layer_norm2" ) if "img_encoder.norm" in name: UpperCAmelCase_ = name.replace("img_encoder.norm" , "vision_model.layernorm" ) # text encoder if "text_encoder.token_embedding" in name: UpperCAmelCase_ = name.replace("text_encoder.token_embedding" , "text_model.embeddings.token_embedding" ) if "text_encoder.positional_embedding" in name: UpperCAmelCase_ = name.replace("text_encoder.positional_embedding" , "text_model.embeddings.position_embedding.weight" ) if "text_encoder.transformer.resblocks." in name: UpperCAmelCase_ = name.replace("text_encoder.transformer.resblocks." , "text_model.encoder.layers." ) if "ln_1" in name: UpperCAmelCase_ = name.replace("ln_1" , "layer_norm1" ) if "ln_2" in name: UpperCAmelCase_ = name.replace("ln_2" , "layer_norm2" ) if "c_fc" in name: UpperCAmelCase_ = name.replace("c_fc" , "fc1" ) if "c_proj" in name: UpperCAmelCase_ = name.replace("c_proj" , "fc2" ) if "text_encoder" in name: UpperCAmelCase_ = name.replace("text_encoder" , "text_model" ) if "ln_final" in name: UpperCAmelCase_ = name.replace("ln_final" , "final_layer_norm" ) # projection layers if "img_projector.linear_hidden." in name: UpperCAmelCase_ = name.replace("img_projector.linear_hidden." , "visual_projection." ) if "img_projector.linear_out." in name: UpperCAmelCase_ = name.replace("img_projector.linear_out." , "visual_projection.3." ) if "text_projector.linear_hidden" in name: UpperCAmelCase_ = name.replace("text_projector.linear_hidden" , "text_projection" ) if "text_projector.linear_out" in name: UpperCAmelCase_ = name.replace("text_projector.linear_out" , "text_projection.3" ) return name def lowerCAmelCase_ ( snake_case_ : Tuple , snake_case_ : List[str] ) -> Union[str, Any]: '''simple docstring''' for key in orig_state_dict.copy().keys(): UpperCAmelCase_ = orig_state_dict.pop(snake_case_ ) if "qkv" in key: # weights and biases of the key, value and query projections of vision encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors UpperCAmelCase_ = key.split("." ) UpperCAmelCase_ , UpperCAmelCase_ = int(key_split[2] ), int(key_split[4] ) UpperCAmelCase_ = config.vision_config.hidden_size if "weight" in key: UpperCAmelCase_ = val[:dim, :] UpperCAmelCase_ = val[dim : dim * 2, :] UpperCAmelCase_ = val[-dim:, :] else: UpperCAmelCase_ = val[:dim] UpperCAmelCase_ = val[dim : dim * 2] UpperCAmelCase_ = val[-dim:] elif "in_proj" in key: # weights and biases of the key, value and query projections of text encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors UpperCAmelCase_ = key.split("." ) UpperCAmelCase_ = int(key_split[3] ) UpperCAmelCase_ = config.text_config.hidden_size if "weight" in key: UpperCAmelCase_ = val[:dim, :] UpperCAmelCase_ = val[ dim : dim * 2, : ] UpperCAmelCase_ = val[-dim:, :] else: UpperCAmelCase_ = val[:dim] UpperCAmelCase_ = val[dim : dim * 2] UpperCAmelCase_ = val[-dim:] else: UpperCAmelCase_ = rename_key(snake_case_ ) # squeeze if necessary if ( "text_projection.0" in new_name or "text_projection.3" in new_name or "visual_projection.0" in new_name or "visual_projection.3" in new_name ): UpperCAmelCase_ = val.squeeze_() else: UpperCAmelCase_ = val return orig_state_dict def lowerCAmelCase_ ( ) -> Any: '''simple docstring''' UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = Image.open(requests.get(snake_case_ , stream=snake_case_ ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : List[Any] , snake_case_ : Optional[Any]="groupvit-gcc-yfcc" , snake_case_ : List[str]=False ) -> Any: '''simple docstring''' UpperCAmelCase_ = GroupViTConfig() UpperCAmelCase_ = GroupViTModel(snake_case_ ).eval() UpperCAmelCase_ = torch.load(snake_case_ , map_location="cpu" )["model"] UpperCAmelCase_ = convert_state_dict(snake_case_ , snake_case_ ) UpperCAmelCase_ , UpperCAmelCase_ = model.load_state_dict(snake_case_ , strict=snake_case_ ) assert missing_keys == ["text_model.embeddings.position_ids"] assert (unexpected_keys == ["multi_label_logit_scale"]) or (len(snake_case_ ) == 0) # verify result UpperCAmelCase_ = CLIPProcessor.from_pretrained("openai/clip-vit-base-patch32" ) UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = processor(text=["a photo of a cat", "a photo of a dog"] , images=snake_case_ , padding=snake_case_ , return_tensors="pt" ) with torch.no_grad(): UpperCAmelCase_ = model(**snake_case_ ) if model_name == "groupvit-gcc-yfcc": UpperCAmelCase_ = torch.tensor([[13.3523, 6.3629]] ) elif model_name == "groupvit-gcc-redcaps": UpperCAmelCase_ = torch.tensor([[16.1873, 8.6230]] ) else: raise ValueError(f"""Model name {model_name} not supported.""" ) assert torch.allclose(outputs.logits_per_image , snake_case_ , atol=1E-3 ) processor.save_pretrained(snake_case_ ) model.save_pretrained(snake_case_ ) print("Successfully saved processor and model to" , snake_case_ ) if push_to_hub: print("Pushing to the hub..." ) processor.push_to_hub(snake_case_ , organization="nielsr" ) model.push_to_hub(snake_case_ , organization="nielsr" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: Dict =argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to dump the processor and PyTorch model.' ) parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to GroupViT checkpoint') parser.add_argument( '--model_name', default='groupvit-gccy-fcc', type=str, help='Name of the model. Expecting either \'groupvit-gcc-yfcc\' or \'groupvit-gcc-redcaps\'', ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model and processor to the 🤗 hub using the provided `model_name`.', ) SCREAMING_SNAKE_CASE_: Optional[int] =parser.parse_args() convert_groupvit_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
1
"""simple docstring""" import torch from diffusers import DiffusionPipeline class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" super().__init__() self.register_modules(unet=snake_case_ , scheduler=snake_case_ ) def __call__( self ): """simple docstring""" A_ : Optional[Any] = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) A_ : List[str] = 1 A_ : List[str] = self.unet(snake_case_ , snake_case_ ).sample A_ : Optional[int] = self.scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample A_ : List[Any] = scheduler_output - scheduler_output + torch.ones_like(snake_case_ ) return result
286
0
'''simple docstring''' from __future__ import annotations def _SCREAMING_SNAKE_CASE (A , A , A ) -> float: """simple docstring""" if days_between_payments <= 0: raise ValueError('''days_between_payments must be > 0''' ) if daily_interest_rate < 0: raise ValueError('''daily_interest_rate must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return principal * daily_interest_rate * days_between_payments def _SCREAMING_SNAKE_CASE (A , A , A , ) -> float: """simple docstring""" if number_of_compounding_periods <= 0: raise ValueError('''number_of_compounding_periods must be > 0''' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('''nominal_annual_interest_rate_percentage must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _SCREAMING_SNAKE_CASE (A , A , A , ) -> float: """simple docstring""" if number_of_years <= 0: raise ValueError('''number_of_years must be > 0''' ) if nominal_annual_percentage_rate < 0: raise ValueError('''nominal_annual_percentage_rate must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return compound_interest( A , nominal_annual_percentage_rate / 365 , number_of_years * 365 ) if __name__ == "__main__": import doctest doctest.testmod()
2
"""simple docstring""" from heapq import heappop, heappush import numpy as np def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" A_ , A_ : List[str] = grid.shape A_ : Optional[int] = [-1, 1, 0, 0] A_ : str = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] A_ , A_ : List[Any] = [(0, source)], set() A_ : Optional[Any] = np.full((rows, cols) , np.inf ) A_ : int = 0 A_ : Optional[int] = np.empty((rows, cols) , dtype=_UpperCAmelCase ) A_ : Optional[int] = None while queue: ((A_) , (A_)) : str = heappop(_UpperCAmelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: A_ : int = [] while (x, y) != source: path.append((x, y) ) A_ , A_ : List[Any] = predecessors[x, y] path.append(_UpperCAmelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(_UpperCAmelCase ) ): A_ , A_ : Tuple = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: A_ : Union[str, Any] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(_UpperCAmelCase , (dist + 1, (nx, ny)) ) A_ : Optional[Any] = dist + 1 A_ : Optional[Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
286
0
'''simple docstring''' lowercase : Optional[int] = {str(digit): digit**5 for digit in range(10)} def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(snake_case__ ) ) def lowerCAmelCase_ ( ): '''simple docstring''' return sum( number for number in range(1000 , 100_0000 ) if number == digits_fifth_powers_sum(snake_case__ ) ) if __name__ == "__main__": print(solution())
3
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase_ : Optional[Any] = { 'huggingface/informer-tourism-monthly': ( 'https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json' ), # See all Informer models at https://huggingface.co/models?filter=informer } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """informer""" lowercase_ : str = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self , snake_case_ = None , snake_case_ = None , snake_case_ = "student_t" , snake_case_ = "nll" , snake_case_ = 1 , snake_case_ = None , snake_case_ = "mean" , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = 6_4 , snake_case_ = 3_2 , snake_case_ = 3_2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = True , snake_case_ = "gelu" , snake_case_ = 0.05 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 1_0_0 , snake_case_ = 0.02 , snake_case_=True , snake_case_ = "prob" , snake_case_ = 5 , snake_case_ = True , **snake_case_ , ): """simple docstring""" A_ : str = prediction_length A_ : List[Any] = context_length or prediction_length A_ : str = distribution_output A_ : Dict = loss A_ : Any = input_size A_ : Union[str, Any] = num_time_features A_ : Optional[Any] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] A_ : List[Any] = scaling A_ : Tuple = num_dynamic_real_features A_ : Any = num_static_real_features A_ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) A_ : Optional[int] = cardinality else: A_ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) A_ : Any = embedding_dimension else: A_ : Optional[Any] = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] A_ : int = num_parallel_samples # Transformer architecture configuration A_ : str = input_size * len(self.lags_sequence ) + self._number_of_features A_ : List[Any] = d_model A_ : Dict = encoder_attention_heads A_ : Dict = decoder_attention_heads A_ : List[Any] = encoder_ffn_dim A_ : Union[str, Any] = decoder_ffn_dim A_ : int = encoder_layers A_ : Any = decoder_layers A_ : List[Any] = dropout A_ : str = attention_dropout A_ : Tuple = activation_dropout A_ : List[str] = encoder_layerdrop A_ : List[str] = decoder_layerdrop A_ : str = activation_function A_ : Optional[int] = init_std A_ : List[Any] = use_cache # Informer A_ : Tuple = attention_type A_ : List[Any] = sampling_factor A_ : Optional[int] = distil super().__init__(is_encoder_decoder=snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
286
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : List[str] = '''naver-clova-ix/donut-base-finetuned-docvqa''' lowerCamelCase : List[Any] = ( '''This is a tool that answers a question about an document (pdf). It takes an input named `document` which ''' '''should be the document containing the information, as well as a `question` that is the question about the ''' '''document. It returns a text that contains the answer to the question.''' ) lowerCamelCase : Union[str, Any] = '''document_qa''' lowerCamelCase : List[str] = AutoProcessor lowerCamelCase : List[str] = VisionEncoderDecoderModel lowerCamelCase : Union[str, Any] = ['''image''', '''text'''] lowerCamelCase : Any = ['''text'''] def __init__( self : Optional[int] , *UpperCAmelCase__ : int , **UpperCAmelCase__ : Optional[Any] ) -> Optional[Any]: if not is_vision_available(): raise ValueError('Pillow must be installed to use the DocumentQuestionAnsweringTool.' ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[Any] , UpperCAmelCase__ : "Image" , UpperCAmelCase__ : str ) -> Dict: lowerCAmelCase = '<s_docvqa><s_question>{user_input}</s_question><s_answer>' lowerCAmelCase = task_prompt.replace('{user_input}' , UpperCAmelCase__ ) lowerCAmelCase = self.pre_processor.tokenizer( UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ , return_tensors='pt' ).input_ids lowerCAmelCase = self.pre_processor(UpperCAmelCase__ , return_tensors='pt' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : Optional[int] ) -> int: return self.model.generate( inputs['pixel_values'].to(self.device ) , decoder_input_ids=inputs['decoder_input_ids'].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=UpperCAmelCase__ , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=UpperCAmelCase__ , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=UpperCAmelCase__ , ).sequences def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : Tuple ) -> Dict: lowerCAmelCase = self.pre_processor.batch_decode(UpperCAmelCase__ )[0] lowerCAmelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '' ) lowerCAmelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '' ) lowerCAmelCase = re.sub(R'<.*?>' , '' , UpperCAmelCase__ , count=1 ).strip() # remove first task start token lowerCAmelCase = self.pre_processor.tokenajson(UpperCAmelCase__ ) return sequence["answer"]
4
"""simple docstring""" import os def UpperCAmelCase__ ( ): """simple docstring""" A_ : Any = os.path.join(os.path.dirname(_UpperCAmelCase ) , 'num.txt' ) with open(_UpperCAmelCase ) as file_hand: return str(sum(int(_UpperCAmelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
286
0
import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## UpperCAmelCase__ = 16 UpperCAmelCase__ = 32 def UpperCAmelCase_ ( __snake_case , __snake_case = 16 ) -> Any: """simple docstring""" _lowercase =AutoTokenizer.from_pretrained('''bert-base-cased''' ) _lowercase =load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__snake_case ): # max_length=None => use the model max length (it's actually the default) _lowercase =tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__snake_case , max_length=__snake_case ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _lowercase =datasets.map( __snake_case , batched=__snake_case , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowercase =tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__snake_case ): # On TPU it's best to pad everything to the same length or training will be very slow. _lowercase =128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _lowercase =16 elif accelerator.mixed_precision != "no": _lowercase =8 else: _lowercase =None return tokenizer.pad( __snake_case , padding='''longest''' , max_length=__snake_case , pad_to_multiple_of=__snake_case , return_tensors='''pt''' , ) # Instantiate dataloaders. _lowercase =DataLoader( tokenized_datasets['''train'''] , shuffle=__snake_case , collate_fn=__snake_case , batch_size=__snake_case , drop_last=__snake_case ) _lowercase =DataLoader( tokenized_datasets['''validation'''] , shuffle=__snake_case , collate_fn=__snake_case , batch_size=__snake_case , drop_last=(accelerator.mixed_precision == '''fp8''') , ) return train_dataloader, eval_dataloader def UpperCAmelCase_ ( __snake_case , __snake_case ) -> List[str]: """simple docstring""" _lowercase =Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowercase =config['''lr'''] _lowercase =int(config['''num_epochs'''] ) _lowercase =int(config['''seed'''] ) _lowercase =int(config['''batch_size'''] ) _lowercase =evaluate.load('''glue''' , '''mrpc''' ) # If the batch size is too big we use gradient accumulation _lowercase =1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _lowercase =batch_size // MAX_GPU_BATCH_SIZE _lowercase =MAX_GPU_BATCH_SIZE set_seed(__snake_case ) _lowercase , _lowercase =get_dataloaders(__snake_case , __snake_case ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowercase =AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__snake_case ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _lowercase =model.to(accelerator.device ) # Instantiate optimizer _lowercase =AdamW(params=model.parameters() , lr=__snake_case ) # Instantiate scheduler _lowercase =get_linear_schedule_with_warmup( optimizer=__snake_case , num_warmup_steps=100 , num_training_steps=(len(__snake_case ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowercase , _lowercase , _lowercase , _lowercase , _lowercase =accelerator.prepare( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) # Now we train the model for epoch in range(__snake_case ): model.train() for step, batch in enumerate(__snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _lowercase =model(**__snake_case ) _lowercase =outputs.loss _lowercase =loss / gradient_accumulation_steps accelerator.backward(__snake_case ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowercase =model(**__snake_case ) _lowercase =outputs.logits.argmax(dim=-1 ) _lowercase , _lowercase =accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__snake_case , references=__snake_case , ) _lowercase =metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __snake_case ) def UpperCAmelCase_ ( ) -> List[str]: """simple docstring""" _lowercase =argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__snake_case , default=__snake_case , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) _lowercase =parser.parse_args() _lowercase ={'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__snake_case , __snake_case ) if __name__ == "__main__": main()
5
"""simple docstring""" import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger lowerCamelCase_ : Dict = get_logger(__name__) lowerCamelCase_ : List[str] = r'\n Args:\n input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam\n search or log softmax for each vocabulary token when using beam search\n kwargs (`Dict[str, Any]`, *optional*):\n Additional logits processor specific kwargs.\n\n Return:\n `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores.\n\n' class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ): """simple docstring""" for processor in self: A_ : Tuple = inspect.signature(processor.__call__ ).parameters if len(snake_case_ ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( F"""Make sure that all the required parameters: {list(function_args.keys() )} for """ F"""{processor.__class__} are passed to the logits processor.""" ) A_ : Tuple = processor(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) else: A_ : Optional[Any] = processor(snake_case_ , snake_case_ , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or not (temperature > 0): raise ValueError(F"""`temperature` has to be a strictly positive float, but is {temperature}""" ) A_ : Optional[int] = temperature def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = scores / self.temperature return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or (top_p < 0 or top_p > 1.0): raise ValueError(F"""`top_p` has to be a float > 0 and < 1, but is {top_p}""" ) if not isinstance(snake_case_ , snake_case_ ) or (min_tokens_to_keep < 1): raise ValueError(F"""`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}""" ) A_ : str = top_p A_ : Union[str, Any] = filter_value A_ : int = min_tokens_to_keep def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : Tuple = lax.top_k(snake_case_ , scores.shape[-1] ) A_ : List[Any] = jnp.full_like(snake_case_ , self.filter_value ) A_ : List[str] = jax.nn.softmax(snake_case_ , axis=-1 ).cumsum(axis=-1 ) A_ : Optional[int] = cumulative_probs < self.top_p # include the token that is higher than top_p as well A_ : Union[str, Any] = jnp.roll(snake_case_ , 1 ) score_mask |= score_mask.at[:, 0].set(snake_case_ ) # min tokens to keep A_ : int = score_mask.at[:, : self.min_tokens_to_keep].set(snake_case_ ) A_ : Optional[Any] = jnp.where(snake_case_ , snake_case_ , snake_case_ ) A_ : List[Any] = jax.lax.sort_key_val(snake_case_ , snake_case_ )[-1] return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or top_k <= 0: raise ValueError(F"""`top_k` has to be a strictly positive integer, but is {top_k}""" ) A_ : str = max(snake_case_ , snake_case_ ) A_ : Union[str, Any] = filter_value def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : int = scores.shape A_ : Tuple = jnp.full(batch_size * vocab_size , self.filter_value ) A_ : Union[str, Any] = min(self.top_k , scores.shape[-1] ) # Safety check A_ , A_ : Dict = lax.top_k(snake_case_ , snake_case_ ) A_ : Optional[int] = jnp.broadcast_to((jnp.arange(snake_case_ ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() A_ : int = topk_scores.flatten() A_ : Any = topk_indices.flatten() + shift A_ : List[str] = next_scores_flat.at[topk_indices_flat].set(snake_case_ ) A_ : Union[str, Any] = next_scores_flat.reshape(snake_case_ , snake_case_ ) return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = bos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Union[str, Any] = 1 - jnp.bool_(cur_len - 1 ) A_ : str = jnp.where(snake_case_ , new_scores.at[:, self.bos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = max_length A_ : Optional[int] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Dict = 1 - jnp.bool_(cur_len - self.max_length + 1 ) A_ : Dict = jnp.where(snake_case_ , new_scores.at[:, self.eos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or min_length < 0: raise ValueError(F"""`min_length` has to be a positive integer, but is {min_length}""" ) if not isinstance(snake_case_ , snake_case_ ) or eos_token_id < 0: raise ValueError(F"""`eos_token_id` has to be a positive integer, but is {eos_token_id}""" ) A_ : Any = min_length A_ : List[Any] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) A_ : Optional[Any] = jnp.where(snake_case_ , scores.at[:, self.eos_token_id].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) A_ : Tuple = begin_index def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = 1 - jnp.bool_(cur_len - self.begin_index ) A_ : int = jnp.where(snake_case_ , scores.at[:, self.begin_suppress_tokens].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = scores.at[..., self.suppress_tokens].set(-float('inf' ) ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Any = dict(snake_case_ ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. A_ : Tuple = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: A_ : Tuple = force_token_array.at[index].set(snake_case_ ) A_ : Any = jnp.intaa(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" def _force_token(snake_case_ ): A_ : List[Any] = scores.shape[0] A_ : Any = self.force_token_array[generation_idx] A_ : Tuple = jnp.ones_like(snake_case_ , dtype=scores.dtype ) * -float('inf' ) A_ : List[Any] = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) A_ : int = lax.dynamic_update_slice(snake_case_ , snake_case_ , (0, current_token) ) return new_scores A_ : int = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(snake_case_ ) , lambda: scores , ) , ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = generate_config.eos_token_id A_ : Optional[int] = generate_config.no_timestamps_token_id A_ : List[str] = generate_config.no_timestamps_token_id + 1 A_ : Any = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(snake_case_ , 'max_initial_timestamp_index' ): A_ : List[Any] = generate_config.max_initial_timestamp_index else: A_ : Any = model_config.vocab_size if self.max_initial_timestamp_index is None: A_ : Optional[Any] = model_config.vocab_size def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[str] = scores.at[:, self.no_timestamps_token_id].set(-float('inf' ) ) def handle_pairs(snake_case_ , snake_case_ ): A_ : Any = jnp.where((cur_len - self.begin_index) >= 1 , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , snake_case_ , ) A_ : Tuple = jnp.where((cur_len - self.begin_index) < 2 , snake_case_ , snake_case_ ) A_ : Any = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , snake_case_ , snake_case_ , ) return jnp.where( snake_case_ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float('inf' ) ) , scores_k.at[: self.eos_token_id].set(-float('inf' ) ) , ) , snake_case_ , ) A_ : Tuple = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) A_ : Optional[Any] = jnp.where(cur_len == self.begin_index , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , snake_case_ , ) A_ : int = self.timestamp_begin + self.max_initial_timestamp_index A_ : List[Any] = jnp.where( snake_case_ , scores.at[:, last_allowed + 1 :].set(-float('inf' ) ) , snake_case_ , ) # if sum of probability over timestamps is above any other token, sample timestamp A_ : Any = jax.nn.log_softmax(snake_case_ , axis=-1 ) def handle_cumulative_probs(snake_case_ , snake_case_ ): A_ : Dict = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) A_ : Optional[Any] = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float('inf' ) ) , snake_case_ , ) A_ : Union[str, Any] = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) return scores
286
0
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __A( a ): def __init__( self , _snake_case , _snake_case ) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM __a = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=_snake_case , scheduler=_snake_case ) @torch.no_grad() def __call__( self , _snake_case = 1 , _snake_case = None , _snake_case = 0.0 , _snake_case = 50 , _snake_case = None , _snake_case = "pil" , _snake_case = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' if isinstance(self.unet.config.sample_size , _snake_case ): __a = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: __a = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(_snake_case , _snake_case ) and len(_snake_case ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_snake_case )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) __a = randn_tensor(_snake_case , generator=_snake_case , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(_snake_case ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output __a = self.unet(_snake_case , _snake_case ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __a = self.scheduler.step( _snake_case , _snake_case , _snake_case , eta=_snake_case , use_clipped_model_output=_snake_case , generator=_snake_case ).prev_sample __a = (image / 2 + 0.5).clamp(0 , 1 ) __a = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __a = self.numpy_to_pil(_snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=_snake_case )
6
"""simple docstring""" import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = R'\w+[.]\d+' A_ : int = re.findall(_UpperCAmelCase , _UpperCAmelCase ) for pat in pats: A_ : Optional[int] = key.replace(_UpperCAmelCase , '_'.join(pat.split('.' ) ) ) return key def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): A_ : Union[str, Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: A_ : List[str] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: A_ : Optional[Any] = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer A_ : int = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: A_ : str = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer A_ : Optional[Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": A_ : Optional[Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight A_ : Tuple = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias A_ : Optional[int] = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=42 ): """simple docstring""" A_ : int = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params A_ : Union[str, Any] = flax_model.init_weights(PRNGKey(_UpperCAmelCase ) ) A_ : Optional[Any] = flatten_dict(_UpperCAmelCase ) A_ : Tuple = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): A_ : Any = rename_key(_UpperCAmelCase ) A_ : List[str] = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters A_ , A_ : Union[str, Any] = rename_key_and_reshape_tensor(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown A_ : str = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase )
286
0
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class A : """simple docstring""" def __init__( self : str,lowercase_ : List[str],lowercase_ : Optional[int]=1_3,lowercase_ : Union[str, Any]=7,lowercase_ : str=True,lowercase_ : Tuple=True,lowercase_ : Any=True,lowercase_ : Optional[Any]=True,lowercase_ : Union[str, Any]=9_9,lowercase_ : Any=6_4,lowercase_ : Optional[int]=3_2,lowercase_ : Union[str, Any]=5,lowercase_ : str=4,lowercase_ : str=3_7,lowercase_ : Optional[Any]="gelu",lowercase_ : str=0.1,lowercase_ : Tuple=0.1,lowercase_ : Optional[int]=5_1_2,lowercase_ : int=1_6,lowercase_ : Dict=2,lowercase_ : Dict=0.02,lowercase_ : List[str]=3,lowercase_ : List[str]=4,lowercase_ : List[str]=None,)-> Dict: '''simple docstring''' A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = embedding_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = scope def snake_case__ ( self : int )-> Dict: '''simple docstring''' A__ = ids_tensor([self.batch_size, self.seq_length],self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None if self.use_token_type_ids: A__ = ids_tensor([self.batch_size, self.seq_length],self.type_vocab_size ) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size],self.type_sequence_label_size ) A__ = ids_tensor([self.batch_size, self.seq_length],self.num_labels ) A__ = ids_tensor([self.batch_size],self.num_choices ) A__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self : List[str] )-> Dict: '''simple docstring''' return MegatronBertConfig( vocab_size=self.vocab_size,hidden_size=self.hidden_size,num_hidden_layers=self.num_hidden_layers,num_attention_heads=self.num_attention_heads,intermediate_size=self.intermediate_size,embedding_size=self.embedding_size,hidden_act=self.hidden_act,hidden_dropout_prob=self.hidden_dropout_prob,attention_probs_dropout_prob=self.attention_probs_dropout_prob,max_position_embeddings=self.max_position_embeddings,type_vocab_size=self.type_vocab_size,is_decoder=lowercase_,initializer_range=self.initializer_range,) def snake_case__ ( self : Union[str, Any],lowercase_ : int,lowercase_ : List[str],lowercase_ : Tuple,lowercase_ : Any,lowercase_ : Any,lowercase_ : int,lowercase_ : Dict )-> Dict: '''simple docstring''' A__ = MegatronBertModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_ ) A__ = model(lowercase_,token_type_ids=lowercase_ ) A__ = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape,(self.batch_size, self.hidden_size) ) def snake_case__ ( self : List[Any],lowercase_ : int,lowercase_ : Tuple,lowercase_ : Optional[Any],lowercase_ : Union[str, Any],lowercase_ : int,lowercase_ : Tuple,lowercase_ : Union[str, Any] )-> Optional[int]: '''simple docstring''' A__ = MegatronBertForMaskedLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : List[Any],lowercase_ : Tuple,lowercase_ : int,lowercase_ : Union[str, Any],lowercase_ : Tuple,lowercase_ : Dict,lowercase_ : List[str],lowercase_ : Union[str, Any] )-> Any: '''simple docstring''' A__ = MegatronBertForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : Optional[Any],lowercase_ : Optional[Any],lowercase_ : Union[str, Any],lowercase_ : Tuple,lowercase_ : str,lowercase_ : str,lowercase_ : Union[str, Any],lowercase_ : List[str] )-> Any: '''simple docstring''' A__ = MegatronBertForNextSentencePrediction(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model( lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_,) self.parent.assertEqual(result.logits.shape,(self.batch_size, 2) ) def snake_case__ ( self : Union[str, Any],lowercase_ : Union[str, Any],lowercase_ : List[str],lowercase_ : Any,lowercase_ : Any,lowercase_ : Optional[Any],lowercase_ : List[Any],lowercase_ : Dict )-> Union[str, Any]: '''simple docstring''' A__ = MegatronBertForPreTraining(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model( lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_,next_sentence_label=lowercase_,) self.parent.assertEqual(result.prediction_logits.shape,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape,(self.batch_size, 2) ) def snake_case__ ( self : Tuple,lowercase_ : str,lowercase_ : Optional[int],lowercase_ : List[Any],lowercase_ : Optional[int],lowercase_ : List[Any],lowercase_ : str,lowercase_ : Dict )-> Dict: '''simple docstring''' A__ = MegatronBertForQuestionAnswering(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model( lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,start_positions=lowercase_,end_positions=lowercase_,) self.parent.assertEqual(result.start_logits.shape,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape,(self.batch_size, self.seq_length) ) def snake_case__ ( self : Union[str, Any],lowercase_ : Union[str, Any],lowercase_ : Dict,lowercase_ : Optional[int],lowercase_ : Any,lowercase_ : Optional[int],lowercase_ : Optional[Any],lowercase_ : Any )-> Optional[Any]: '''simple docstring''' A__ = self.num_labels A__ = MegatronBertForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.num_labels) ) def snake_case__ ( self : Any,lowercase_ : Dict,lowercase_ : Optional[int],lowercase_ : Union[str, Any],lowercase_ : Tuple,lowercase_ : Optional[Any],lowercase_ : str,lowercase_ : Tuple )-> int: '''simple docstring''' A__ = self.num_labels A__ = MegatronBertForTokenClassification(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self : Any,lowercase_ : Any,lowercase_ : Union[str, Any],lowercase_ : Tuple,lowercase_ : int,lowercase_ : Optional[Any],lowercase_ : Dict,lowercase_ : Any )-> Any: '''simple docstring''' A__ = self.num_choices A__ = MegatronBertForMultipleChoice(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = input_ids.unsqueeze(1 ).expand(-1,self.num_choices,-1 ).contiguous() A__ = token_type_ids.unsqueeze(1 ).expand(-1,self.num_choices,-1 ).contiguous() A__ = input_mask.unsqueeze(1 ).expand(-1,self.num_choices,-1 ).contiguous() A__ = model( lowercase_,attention_mask=lowercase_,token_type_ids=lowercase_,labels=lowercase_,) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.num_choices) ) def snake_case__ ( self : List[Any] )-> Any: '''simple docstring''' A__ = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) = config_and_inputs A__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase = ( { 'feature-extraction': MegatronBertModel, 'fill-mask': MegatronBertForMaskedLM, 'question-answering': MegatronBertForQuestionAnswering, 'text-classification': MegatronBertForSequenceClassification, 'text-generation': MegatronBertForCausalLM, 'token-classification': MegatronBertForTokenClassification, 'zero-shot': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase = True # test_resize_embeddings = False lowerCamelCase = False def snake_case__ ( self : List[Any],lowercase_ : Any,lowercase_ : List[Any],lowercase_ : List[str]=False )-> Optional[Any]: '''simple docstring''' A__ = super()._prepare_for_class(lowercase_,lowercase_,return_labels=lowercase_ ) if return_labels: if model_class in get_values(lowercase_ ): A__ = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length),dtype=torch.long,device=lowercase_ ) A__ = torch.zeros( self.model_tester.batch_size,dtype=torch.long,device=lowercase_ ) return inputs_dict def snake_case__ ( self : Optional[Any] )-> int: '''simple docstring''' A__ = MegatronBertModelTester(self ) A__ = ConfigTester(self,config_class=lowercase_,hidden_size=3_7 ) def snake_case__ ( self : int )-> int: '''simple docstring''' self.config_tester.run_common_tests() def snake_case__ ( self : Tuple )-> Optional[int]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*lowercase_ ) def snake_case__ ( self : int )-> str: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*lowercase_ ) def snake_case__ ( self : List[Any] )-> Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*lowercase_ ) def snake_case__ ( self : str )-> Dict: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*lowercase_ ) def snake_case__ ( self : Dict )-> Any: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*lowercase_ ) def snake_case__ ( self : int )-> Optional[int]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*lowercase_ ) def snake_case__ ( self : Optional[Any] )-> str: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*lowercase_ ) def snake_case__ ( self : List[Any] )-> Optional[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*lowercase_ ) def _snake_case( SCREAMING_SNAKE_CASE__ : int ) -> Any: '''simple docstring''' return torch.tensor( SCREAMING_SNAKE_CASE__ , dtype=torch.long , device=SCREAMING_SNAKE_CASE__ , ) lowercase_ = 1e-4 @require_torch @require_sentencepiece @require_tokenizers class A ( unittest.TestCase ): """simple docstring""" @slow @unittest.skip('Model is not available.' ) def snake_case__ ( self : Any )-> Tuple: '''simple docstring''' A__ = 'nvidia/megatron-bert-uncased-345m' if "MYDIR" in os.environ: A__ = os.path.join(os.environ['MYDIR'],lowercase_ ) A__ = MegatronBertModel.from_pretrained(lowercase_ ) model.to(lowercase_ ) model.half() A__ = _long_tensor([[1_0_1, 7_1_1_0, 1_0_0_5, 1_0_5_6, 2_0_2_3, 1_1_3_3_3, 1_7_4_1_3, 1_0_2_9, 1_0_2]] ) with torch.no_grad(): A__ = model(lowercase_ )[0] A__ = torch.Size((1, 9, 1_0_2_4) ) self.assertEqual(output.shape,lowercase_ ) A__ = [-0.6_040, -0.2_517, -0.1_025, 0.3_420, -0.6_758, -0.0_017, -0.1_089, -0.1_990, 0.5_728] for ii in range(3 ): for jj in range(3 ): A__ = output[0, ii, jj] A__ = expected[3 * ii + jj] A__ = 'ii={} jj={} a={} b={}'.format(lowercase_,lowercase_,lowercase_,lowercase_ ) self.assertTrue(math.isclose(lowercase_,lowercase_,rel_tol=lowercase_,abs_tol=lowercase_ ),msg=lowercase_ )
7
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = CustomTokenizer pass
286
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class snake_case_ ( __A ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = "vivit" def __init__( self : int , _UpperCamelCase : Union[str, Any]=2_2_4 , _UpperCamelCase : List[str]=3_2 , _UpperCamelCase : Optional[Any]=[2, 1_6, 1_6] , _UpperCamelCase : Dict=3 , _UpperCamelCase : int=7_6_8 , _UpperCamelCase : Any=1_2 , _UpperCamelCase : Dict=1_2 , _UpperCamelCase : Union[str, Any]=3_0_7_2 , _UpperCamelCase : List[Any]="gelu_fast" , _UpperCamelCase : str=0.0 , _UpperCamelCase : Dict=0.0 , _UpperCamelCase : str=0.02 , _UpperCamelCase : Optional[int]=1e-06 , _UpperCamelCase : Any=True , **_UpperCamelCase : List[Any] , ) ->int: snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = image_size snake_case_ = num_frames snake_case_ = tubelet_size snake_case_ = num_channels snake_case_ = qkv_bias super().__init__(**_UpperCamelCase )
8
"""simple docstring""" from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch lowerCamelCase_ : str = logging.get_logger(__name__) @add_end_docstrings( UpperCAmelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.framework == "tf": A_ : str = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": A_ : List[str] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ) else: raise ValueError('Unsupported framework' ) return masked_index def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = self.get_masked_index(snake_case_ ) A_ : str = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['input_ids'][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_=None , **snake_case_ ): """simple docstring""" if return_tensors is None: A_ : Any = self.framework A_ : Dict = self.tokenizer(snake_case_ , return_tensors=snake_case_ ) self.ensure_exactly_one_mask_token(snake_case_ ) return model_inputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Dict = self.model(**snake_case_ ) A_ : Optional[int] = model_inputs['input_ids'] return model_outputs def lowerCamelCase_ ( self , snake_case_ , snake_case_=5 , snake_case_=None ): """simple docstring""" if target_ids is not None and target_ids.shape[0] < top_k: A_ : str = target_ids.shape[0] A_ : Optional[Any] = model_outputs['input_ids'][0] A_ : List[Any] = model_outputs['logits'] if self.framework == "tf": A_ : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] A_ : Union[str, Any] = outputs.numpy() A_ : Optional[int] = outputs[0, masked_index, :] A_ : Optional[Any] = stable_softmax(snake_case_ , axis=-1 ) if target_ids is not None: A_ : Union[str, Any] = tf.gather_nd(tf.squeeze(snake_case_ , 0 ) , target_ids.reshape(-1 , 1 ) ) A_ : Optional[int] = tf.expand_dims(snake_case_ , 0 ) A_ : Any = tf.math.top_k(snake_case_ , k=snake_case_ ) A_ , A_ : str = topk.values.numpy(), topk.indices.numpy() else: A_ : int = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample A_ : Tuple = outputs[0, masked_index, :] A_ : List[str] = logits.softmax(dim=-1 ) if target_ids is not None: A_ : str = probs[..., target_ids] A_ , A_ : List[str] = probs.topk(snake_case_ ) A_ : List[Any] = [] A_ : int = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): A_ : str = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place A_ : Union[str, Any] = input_ids.numpy().copy() if target_ids is not None: A_ : str = target_ids[p].tolist() A_ : Union[str, Any] = p # Filter padding out: A_ : Any = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back A_ : Any = self.tokenizer.decode(snake_case_ , skip_special_tokens=snake_case_ ) A_ : Any = {'score': v, 'token': p, 'token_str': self.tokenizer.decode([p] ), 'sequence': sequence} row.append(snake_case_ ) result.append(snake_case_ ) if single_mask: return result[0] return result def lowerCamelCase_ ( self , snake_case_ , snake_case_=None ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : List[str] = [targets] try: A_ : Optional[int] = self.tokenizer.get_vocab() except Exception: A_ : int = {} A_ : Tuple = [] for target in targets: A_ : int = vocab.get(snake_case_ , snake_case_ ) if id_ is None: A_ : Tuple = self.tokenizer( snake_case_ , add_special_tokens=snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , max_length=1 , truncation=snake_case_ , )['input_ids'] if len(snake_case_ ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ 'We cannot replace it with anything meaningful, ignoring it' ) continue A_ : str = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) A_ : Tuple = list(set(snake_case_ ) ) if len(snake_case_ ) == 0: raise ValueError('At least one target must be provided when passed.' ) A_ : Optional[Any] = np.array(snake_case_ ) return target_ids def lowerCamelCase_ ( self , snake_case_=None , snake_case_=None ): """simple docstring""" A_ : List[str] = {} if targets is not None: A_ : Any = self.get_target_ids(snake_case_ , snake_case_ ) A_ : Optional[Any] = target_ids if top_k is not None: A_ : int = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , 'The tokenizer does not define a `mask_token`.' ) return {}, {}, postprocess_params def __call__( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : List[str] = super().__call__(snake_case_ , **snake_case_ ) if isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1: return outputs[0] return outputs
286
0
from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split __lowerCAmelCase : List[Any] =datasets.load_iris() __lowerCAmelCase : Tuple =np.array(data['data']) __lowerCAmelCase : Dict =np.array(data['target']) __lowerCAmelCase : List[str] =data['target_names'] __lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase : str =train_test_split(X, y) def _UpperCamelCase ( lowercase__ , lowercase__ ): return np.linalg.norm(np.array(lowercase__ ) - np.array(lowercase__ ) ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__=5 ): __SCREAMING_SNAKE_CASE : Optional[int] = zip(lowercase__ , lowercase__ ) # List of distances of all points from the point to be classified __SCREAMING_SNAKE_CASE : Dict = [] for data_point in data: __SCREAMING_SNAKE_CASE : Tuple = euclidean_distance(data_point[0] , lowercase__ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. __SCREAMING_SNAKE_CASE : int = [i[1] for i in sorted(lowercase__ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified __SCREAMING_SNAKE_CASE : Any = Counter(lowercase__ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
9
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) def lowerCAmelCase_ ( __a ) -> YolosConfig: """simple docstring""" lowerCamelCase__: str =YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: lowerCamelCase__: int =192 lowerCamelCase__: Optional[int] =768 lowerCamelCase__: Any =12 lowerCamelCase__: str =3 lowerCamelCase__: Optional[int] =[800, 1333] lowerCamelCase__: Union[str, Any] =False elif yolos_name == "yolos_s_dWr": lowerCamelCase__: int =330 lowerCamelCase__: Optional[Any] =14 lowerCamelCase__: Any =6 lowerCamelCase__: List[str] =1320 elif "yolos_s" in yolos_name: lowerCamelCase__: List[str] =384 lowerCamelCase__: Union[str, Any] =1536 lowerCamelCase__: List[Any] =12 lowerCamelCase__: Any =6 elif "yolos_b" in yolos_name: lowerCamelCase__: str =[800, 1344] lowerCamelCase__: int =91 lowerCamelCase__: str ="huggingface/label-files" lowerCamelCase__: List[str] ="coco-detection-id2label.json" lowerCamelCase__: Tuple =json.load(open(hf_hub_download(__a , __a , repo_type="dataset" ) , "r" ) ) lowerCamelCase__: Dict ={int(__a ): v for k, v in idalabel.items()} lowerCamelCase__: List[str] =idalabel lowerCamelCase__: int ={v: k for k, v in idalabel.items()} return config def lowerCAmelCase_ ( __a , __a , __a = False ) -> Dict: """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__: Optional[int] =state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) lowerCamelCase__: Dict =state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__: Union[str, Any] =in_proj_weight[: config.hidden_size, :] lowerCamelCase__: str =in_proj_bias[: config.hidden_size] lowerCamelCase__: str =in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__: str =in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__: Optional[int] =in_proj_weight[-config.hidden_size :, :] lowerCamelCase__: List[Any] =in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( __a ) -> str: """simple docstring""" if "backbone" in name: lowerCamelCase__: Optional[Any] =name.replace("backbone" , "vit" ) if "cls_token" in name: lowerCamelCase__: Optional[int] =name.replace("cls_token" , "embeddings.cls_token" ) if "det_token" in name: lowerCamelCase__: str =name.replace("det_token" , "embeddings.detection_tokens" ) if "mid_pos_embed" in name: lowerCamelCase__: Tuple =name.replace("mid_pos_embed" , "encoder.mid_position_embeddings" ) if "pos_embed" in name: lowerCamelCase__: Any =name.replace("pos_embed" , "embeddings.position_embeddings" ) if "patch_embed.proj" in name: lowerCamelCase__: List[Any] =name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "blocks" in name: lowerCamelCase__: Union[str, Any] =name.replace("blocks" , "encoder.layer" ) if "attn.proj" in name: lowerCamelCase__: Any =name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: lowerCamelCase__: Optional[int] =name.replace("attn" , "attention.self" ) if "norm1" in name: lowerCamelCase__: int =name.replace("norm1" , "layernorm_before" ) if "norm2" in name: lowerCamelCase__: int =name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: lowerCamelCase__: List[str] =name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: lowerCamelCase__: Any =name.replace("mlp.fc2" , "output.dense" ) if "class_embed" in name: lowerCamelCase__: Dict =name.replace("class_embed" , "class_labels_classifier" ) if "bbox_embed" in name: lowerCamelCase__: List[str] =name.replace("bbox_embed" , "bbox_predictor" ) if "vit.norm" in name: lowerCamelCase__: Any =name.replace("vit.norm" , "vit.layernorm" ) return name def lowerCAmelCase_ ( __a , __a ) -> dict: """simple docstring""" for key in orig_state_dict.copy().keys(): lowerCamelCase__: Any =orig_state_dict.pop(__a ) if "qkv" in key: lowerCamelCase__: Tuple =key.split("." ) lowerCamelCase__: List[str] =int(key_split[2] ) lowerCamelCase__: Tuple =model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: lowerCamelCase__: int =val[:dim, :] lowerCamelCase__: str =val[ dim : dim * 2, : ] lowerCamelCase__: Any =val[-dim:, :] else: lowerCamelCase__: Tuple =val[:dim] lowerCamelCase__: Optional[Any] =val[dim : dim * 2] lowerCamelCase__: str =val[-dim:] else: lowerCamelCase__: Dict =val return orig_state_dict def lowerCAmelCase_ ( ) -> torch.Tensor: """simple docstring""" lowerCamelCase__: Any ="http://images.cocodataset.org/val2017/000000039769.jpg" lowerCamelCase__: Optional[Any] =Image.open(requests.get(__a , stream=__a ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( __a , __a , __a , __a = False ) -> List[str]: """simple docstring""" lowerCamelCase__: int =get_yolos_config(__a ) # load original state_dict lowerCamelCase__: Optional[int] =torch.load(__a , map_location="cpu" )["model"] # load 🤗 model lowerCamelCase__: int =YolosForObjectDetection(__a ) model.eval() lowerCamelCase__: Union[str, Any] =convert_state_dict(__a , __a ) model.load_state_dict(__a ) # Check outputs on an image, prepared by YolosImageProcessor lowerCamelCase__: Any =800 if yolos_name != "yolos_ti" else 512 lowerCamelCase__: Tuple =YolosImageProcessor(format="coco_detection" , size=__a ) lowerCamelCase__: str =image_processor(images=prepare_img() , return_tensors="pt" ) lowerCamelCase__: Tuple =model(**__a ) lowerCamelCase__ , lowerCamelCase__: List[str] =outputs.logits, outputs.pred_boxes lowerCamelCase__ , lowerCamelCase__: Any =None, None if yolos_name == "yolos_ti": lowerCamelCase__: Optional[Any] =torch.tensor( [[-3_9.5_0_2_2, -1_1.9_8_2_0, -1_7.6_8_8_8], [-2_9.9_5_7_4, -9.9_7_6_9, -1_7.7_6_9_1], [-4_2.3_2_8_1, -2_0.7_2_0_0, -3_0.6_2_9_4]] ) lowerCamelCase__: List[Any] =torch.tensor( [[0.4_0_2_1, 0.0_8_3_6, 0.7_9_7_9], [0.0_1_8_4, 0.2_6_0_9, 0.0_3_6_4], [0.1_7_8_1, 0.2_0_0_4, 0.2_0_9_5]] ) elif yolos_name == "yolos_s_200_pre": lowerCamelCase__: Optional[int] =torch.tensor( [[-2_4.0_2_4_8, -1_0.3_0_2_4, -1_4.8_2_9_0], [-4_2.0_3_9_2, -1_6.8_2_0_0, -2_7.4_3_3_4], [-2_7.2_7_4_3, -1_1.8_1_5_4, -1_8.7_1_4_8]] ) lowerCamelCase__: Any =torch.tensor( [[0.2_5_5_9, 0.5_4_5_5, 0.4_7_0_6], [0.2_9_8_9, 0.7_2_7_9, 0.1_8_7_5], [0.7_7_3_2, 0.4_0_1_7, 0.4_4_6_2]] ) elif yolos_name == "yolos_s_300_pre": lowerCamelCase__: str =torch.tensor( [[-3_6.2_2_2_0, -1_4.4_3_8_5, -2_3.5_4_5_7], [-3_5.6_9_7_0, -1_4.7_5_8_3, -2_1.3_9_3_5], [-3_1.5_9_3_9, -1_3.6_0_4_2, -1_6.8_0_4_9]] ) lowerCamelCase__: Optional[Any] =torch.tensor( [[0.7_6_1_4, 0.2_3_1_6, 0.4_7_2_8], [0.7_1_6_8, 0.4_4_9_5, 0.3_8_5_5], [0.4_9_9_6, 0.1_4_6_6, 0.9_9_9_6]] ) elif yolos_name == "yolos_s_dWr": lowerCamelCase__: str =torch.tensor( [[-4_2.8_6_6_8, -2_4.1_0_4_9, -4_1.1_6_9_0], [-3_4.7_4_5_6, -1_4.1_2_7_4, -2_4.9_1_9_4], [-3_3.7_8_9_8, -1_2.1_9_4_6, -2_5.6_4_9_5]] ) lowerCamelCase__: Union[str, Any] =torch.tensor( [[0.5_5_8_7, 0.2_7_7_3, 0.0_6_0_5], [0.5_0_0_4, 0.3_0_1_4, 0.9_9_9_4], [0.4_9_9_9, 0.1_5_4_8, 0.9_9_9_4]] ) elif yolos_name == "yolos_base": lowerCamelCase__: Tuple =torch.tensor( [[-4_0.6_0_6_4, -2_4.3_0_8_4, -3_2.6_4_4_7], [-5_5.1_9_9_0, -3_0.7_7_1_9, -3_5.5_8_7_7], [-5_1.4_3_1_1, -3_3.3_5_0_7, -3_5.6_4_6_2]] ) lowerCamelCase__: Optional[int] =torch.tensor( [[0.5_5_5_5, 0.2_7_9_4, 0.0_6_5_5], [0.9_0_4_9, 0.2_6_6_4, 0.1_8_9_4], [0.9_1_8_3, 0.1_9_8_4, 0.1_6_3_5]] ) else: raise ValueError(F"""Unknown yolos_name: {yolos_name}""" ) assert torch.allclose(logits[0, :3, :3] , __a , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , __a , atol=1e-4 ) Path(__a ).mkdir(exist_ok=__a ) print(F"""Saving model {yolos_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__a ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__a ) if push_to_hub: lowerCamelCase__: Any ={ "yolos_ti": "yolos-tiny", "yolos_s_200_pre": "yolos-small", "yolos_s_300_pre": "yolos-small-300", "yolos_s_dWr": "yolos-small-dwr", "yolos_base": "yolos-base", } print("Pushing to the hub..." ) lowerCamelCase__: Optional[int] =model_mapping[yolos_name] image_processor.push_to_hub(__a , organization="hustvl" ) model.push_to_hub(__a , organization="hustvl" ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--yolos_name", default="yolos_s_200_pre", type=str, help=( "Name of the YOLOS model you'd like to convert. Should be one of 'yolos_ti', 'yolos_s_200_pre'," " 'yolos_s_300_pre', 'yolos_s_dWr', 'yolos_base'." ), ) parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original state dict (.pth file)." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __A = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
10
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) A_ : Any = hex_num[0] == '-' if is_negative: A_ : Optional[Any] = hex_num[1:] try: A_ : Tuple = int(_UpperCAmelCase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) A_ : Union[str, Any] = '' while int_num > 0: A_ : Optional[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
286
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) lowerCAmelCase__ = { 'configuration_speecht5': [ 'SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP', 'SpeechT5Config', 'SpeechT5HifiGanConfig', ], 'feature_extraction_speecht5': ['SpeechT5FeatureExtractor'], 'processing_speecht5': ['SpeechT5Processor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['SpeechT5Tokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ 'SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'SpeechT5ForSpeechToText', 'SpeechT5ForSpeechToSpeech', 'SpeechT5ForTextToSpeech', 'SpeechT5Model', 'SpeechT5PreTrainedModel', 'SpeechT5HifiGan', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
11
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase_ = { 'configuration_maskformer': ['MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MaskFormerConfig'], 'configuration_maskformer_swin': ['MaskFormerSwinConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['MaskFormerFeatureExtractor'] UpperCAmelCase_ = ['MaskFormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'MaskFormerForInstanceSegmentation', 'MaskFormerModel', 'MaskFormerPreTrainedModel', ] UpperCAmelCase_ = [ 'MaskFormerSwinBackbone', 'MaskFormerSwinModel', 'MaskFormerSwinPreTrainedModel', ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure)
12
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Any = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """xmod""" def __init__( self , snake_case_=3_0_5_2_2 , snake_case_=7_6_8 , snake_case_=1_2 , snake_case_=1_2 , snake_case_=3_0_7_2 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , snake_case_=False , snake_case_=2 , snake_case_=False , snake_case_=True , snake_case_=True , snake_case_=("en_XX",) , snake_case_=None , **snake_case_ , ): """simple docstring""" super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) A_ : Union[str, Any] = vocab_size A_ : Any = hidden_size A_ : List[str] = num_hidden_layers A_ : Tuple = num_attention_heads A_ : int = hidden_act A_ : Any = intermediate_size A_ : Any = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Union[str, Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : List[str] = initializer_range A_ : Any = layer_norm_eps A_ : Optional[Any] = position_embedding_type A_ : int = use_cache A_ : Dict = classifier_dropout A_ : int = pre_norm A_ : Optional[Any] = adapter_reduction_factor A_ : List[Any] = adapter_layer_norm A_ : int = adapter_reuse_layer_norm A_ : Dict = ln_before_adapter A_ : List[str] = list(snake_case_ ) A_ : Union[str, Any] = default_language class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @property def lowerCamelCase_ ( self ): """simple docstring""" if self.task == "multiple-choice": A_ : Dict = {0: 'batch', 1: 'choice', 2: 'sequence'} else: A_ : int = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
286
0
def A_ ( _UpperCAmelCase = 10_00 ): return sum(e for e in range(3 , _UpperCAmelCase ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(f'''{solution() = }''')
13
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
0
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' UpperCAmelCase__ = 42 class UpperCamelCase_ ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self : Optional[int] , UpperCAmelCase__ : int = 16 , UpperCAmelCase__ : int = 88 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : int = 32 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : str = "geglu" , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : bool = True , ) ->Union[str, Any]: '''simple docstring''' super().__init__() A__ = num_attention_heads A__ = attention_head_dim A__ = num_attention_heads * attention_head_dim A__ = in_channels A__ = torch.nn.GroupNorm(num_groups=UpperCAmelCase__ , num_channels=UpperCAmelCase__ , eps=1e-6 , affine=UpperCAmelCase__) A__ = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__) # 3. Define transformers blocks A__ = nn.ModuleList( [ BasicTransformerBlock( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , dropout=UpperCAmelCase__ , cross_attention_dim=UpperCAmelCase__ , activation_fn=UpperCAmelCase__ , attention_bias=UpperCAmelCase__ , double_self_attention=UpperCAmelCase__ , norm_elementwise_affine=UpperCAmelCase__ , ) for d in range(UpperCAmelCase__) ]) A__ = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : int=None , UpperCAmelCase__ : bool = True , ) ->List[str]: '''simple docstring''' A__ , A__ , A__ , A__ = hidden_states.shape A__ = batch_frames // num_frames A__ = hidden_states A__ = hidden_states[None, :].reshape(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) A__ = hidden_states.permute(0 , 2 , 1 , 3 , 4) A__ = self.norm(UpperCAmelCase__) A__ = hidden_states.permute(0 , 3 , 4 , 2 , 1).reshape(batch_size * height * width , UpperCAmelCase__ , UpperCAmelCase__) A__ = self.proj_in(UpperCAmelCase__) # 2. Blocks for block in self.transformer_blocks: A__ = block( UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , timestep=UpperCAmelCase__ , cross_attention_kwargs=UpperCAmelCase__ , class_labels=UpperCAmelCase__ , ) # 3. Output A__ = self.proj_out(UpperCAmelCase__) A__ = ( hidden_states[None, None, :] .reshape(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) .permute(0 , 3 , 4 , 1 , 2) .contiguous() ) A__ = hidden_states.reshape(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) A__ = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=UpperCAmelCase__)
14
"""simple docstring""" from copy import deepcopy class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None , snake_case_ = None ): """simple docstring""" if arr is None and size is not None: A_ : Union[str, Any] = size A_ : List[str] = [0] * size elif arr is not None: self.init(snake_case_ ) else: raise ValueError('Either arr or size must be specified' ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = len(snake_case_ ) A_ : Optional[int] = deepcopy(snake_case_ ) for i in range(1 , self.size ): A_ : Optional[Any] = self.next_(snake_case_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): A_ : Optional[int] = self.next_(snake_case_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index + (index & (-index)) @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index - (index & (-index)) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value A_ : List[str] = self.next_(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" self.add(snake_case_ , value - self.get(snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if right == 0: return 0 A_ : Any = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] A_ : Tuple = self.prev(snake_case_ ) return result def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" return self.prefix(snake_case_ ) - self.prefix(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.query(snake_case_ , index + 1 ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" value -= self.tree[0] if value < 0: return -1 A_ : List[Any] = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 A_ : Tuple = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
286
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def UpperCAmelCase ( ) -> Any: """simple docstring""" __A = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=a_ ) __A = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=a_ ) env_command_parser(subparsers=a_ ) launch_command_parser(subparsers=a_ ) tpu_command_parser(subparsers=a_ ) test_command_parser(subparsers=a_ ) # Let's go __A = parser.parse_args() if not hasattr(a_ , "func" ): parser.print_help() exit(1 ) # Run args.func(a_ ) if __name__ == "__main__": main()
15
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , snake_case_ = 7_6_8 , ): """simple docstring""" super().__init__() A_ : Optional[int] = nn.Parameter(torch.zeros(1 , snake_case_ ) ) A_ : Optional[int] = nn.Parameter(torch.ones(1 , snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ = None , snake_case_ = None , ): """simple docstring""" A_ : str = nn.Parameter(self.mean.to(snake_case_ ).to(snake_case_ ) ) A_ : Optional[int] = nn.Parameter(self.std.to(snake_case_ ).to(snake_case_ ) ) return self def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Tuple = (embeds - self.mean) * 1.0 / self.std return embeds def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = (embeds * self.std) + self.mean return embeds
286
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowerCAmelCase_ = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
16
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCamelCase_ : Any = HfArgumentParser(InitializationArguments) lowerCamelCase_ : Union[str, Any] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowerCamelCase_ : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowerCamelCase_ : Tuple = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) lowerCamelCase_ : int = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowerCamelCase_ : Any = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
286
0
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self : List[str] ): __lowercase = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small", return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) __lowercase = AutoTokenizer.from_pretrained("google/mt5-small" ) __lowercase = tokenizer("Hello there", return_tensors="pt" ).input_ids __lowercase = tokenizer("Hi I am", return_tensors="pt" ).input_ids __lowercase = model(input_ids.to(UpperCAmelCase__ ), labels=labels.to(UpperCAmelCase__ ) ).loss __lowercase = -(labels.shape[-1] * loss.item()) __lowercase = -84.9_127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
17
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser lowerCamelCase_ : Any = re.compile(r'\s+') def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return {"hash": hashlib.mda(re.sub(_UpperCAmelCase , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [len(_UpperCAmelCase ) for line in example['content'].splitlines()] return {"line_mean": np.mean(_UpperCAmelCase ), "line_max": max(_UpperCAmelCase )} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 ): """simple docstring""" A_ : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] A_ : List[str] = example['content'].splitlines() for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 , _UpperCAmelCase=0.05 ): """simple docstring""" A_ : Any = ['unit tests', 'test file', 'configuration file'] A_ : Dict = example['content'].splitlines() A_ : List[Any] = 0 A_ : str = 0 # first test for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test A_ : Tuple = example['content'].count('\n' ) A_ : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = ['def ', 'class ', 'for ', 'while '] A_ : Tuple = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=4 ): """simple docstring""" A_ : Union[str, Any] = example['content'].splitlines() A_ : Any = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = tokenizer(example['content'] , truncation=_UpperCAmelCase )['input_ids'] A_ : Dict = len(example['content'] ) / len(_UpperCAmelCase ) return {"ratio": ratio} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = {} results.update(get_hash(_UpperCAmelCase ) ) results.update(line_stats(_UpperCAmelCase ) ) results.update(alpha_stats(_UpperCAmelCase ) ) results.update(char_token_ratio(_UpperCAmelCase ) ) results.update(is_autogenerated(_UpperCAmelCase ) ) results.update(is_config_or_test(_UpperCAmelCase ) ) results.update(has_no_keywords(_UpperCAmelCase ) ) results.update(has_few_assignments(_UpperCAmelCase ) ) return results def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if not check_uniques(_UpperCAmelCase , _UpperCAmelCase ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'rb' ) as f_in: with gzip.open(str(_UpperCAmelCase ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) os.unlink(_UpperCAmelCase ) # Settings lowerCamelCase_ : Optional[int] = HfArgumentParser(PreprocessingArguments) lowerCamelCase_ : Optional[Any] = parser.parse_args() if args.num_workers is None: lowerCamelCase_ : int = multiprocessing.cpu_count() lowerCamelCase_ : Tuple = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset lowerCamelCase_ : Tuple = time.time() lowerCamelCase_ : Tuple = load_dataset(args.dataset_name, split='train') print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing lowerCamelCase_ : List[str] = time.time() lowerCamelCase_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes lowerCamelCase_ : int = set(ds.unique('hash')) lowerCamelCase_ : Union[str, Any] = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics lowerCamelCase_ : Optional[int] = time.time() lowerCamelCase_ : Tuple = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: lowerCamelCase_ : Union[str, Any] = time.time() lowerCamelCase_ , lowerCamelCase_ : str = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file lowerCamelCase_ : Tuple = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) lowerCamelCase_ : Optional[Any] = output_dir / 'data' data_dir.mkdir(exist_ok=True) lowerCamelCase_ : List[str] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): lowerCamelCase_ : Optional[int] = str(data_dir / F"file-{file_number+1:012}.json") lowerCamelCase_ : List[str] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
286
0
import csv import tweepy # Twitter API credentials __lowerCamelCase : Dict = '''''' __lowerCamelCase : Union[str, Any] = '''''' __lowerCamelCase : Dict = '''''' __lowerCamelCase : List[Any] = '''''' def _snake_case ( lowerCAmelCase : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = tweepy.OAuthHandler(lowerCAmelCase , lowerCAmelCase ) auth.set_access_token(lowerCAmelCase , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = tweepy.API(lowerCAmelCase ) # initialize a list to hold all the tweepy Tweets SCREAMING_SNAKE_CASE_ : int = [] # make initial request for most recent tweets (200 is the maximum allowed count) SCREAMING_SNAKE_CASE_ : List[Any] = api.user_timeline(screen_name=lowerCAmelCase , count=2_0_0 ) # save most recent tweets alltweets.extend(lowerCAmelCase ) # save the id of the oldest tweet less one SCREAMING_SNAKE_CASE_ : List[Any] = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(lowerCAmelCase ) > 0: print(f'getting tweets before {oldest}' ) # all subsequent requests use the max_id param to prevent duplicates SCREAMING_SNAKE_CASE_ : int = api.user_timeline( screen_name=lowerCAmelCase , count=2_0_0 , max_id=lowerCAmelCase ) # save most recent tweets alltweets.extend(lowerCAmelCase ) # update the id of the oldest tweet less one SCREAMING_SNAKE_CASE_ : str = alltweets[-1].id - 1 print(f'...{len(lowerCAmelCase )} tweets downloaded so far' ) # transform the tweepy tweets into a 2D array that will populate the csv SCREAMING_SNAKE_CASE_ : Dict = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(f'new_{screen_name}_tweets.csv' , "w" ) as f: SCREAMING_SNAKE_CASE_ : Union[str, Any] = csv.writer(lowerCAmelCase ) writer.writerow(["id", "created_at", "text"] ) writer.writerows(lowerCAmelCase ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('''FirePing32''')
18
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : Optional[Any] = [] # fmt: off # stem: rename_keys.append(('cls_token', 'vit.embeddings.cls_token') ) rename_keys.append(('pos_embed', 'vit.embeddings.position_embeddings') ) rename_keys.append(('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias') ) # backbone rename_keys.append(('patch_embed.backbone.stem.conv.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.bias', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A_ : List[str] = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) # fmt: on return rename_keys def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: A_ : List[str] = '' else: A_ : Dict = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ : str = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) A_ : List[Any] = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A_ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A_ : Tuple = in_proj_bias[: config.hidden_size] A_ : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A_ : Tuple = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = dct.pop(_UpperCAmelCase ) A_ : Optional[int] = val def UpperCAmelCase__ ( ): """simple docstring""" A_ : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' A_ : int = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : List[Any] = BitConfig( global_padding='same' , layer_type='bottleneck' , depths=(3, 4, 9) , out_features=['stage3'] , embedding_dynamic_padding=_UpperCAmelCase , ) A_ : Optional[int] = ViTHybridConfig(backbone_config=_UpperCAmelCase , image_size=384 , num_labels=1000 ) A_ : Union[str, Any] = False # load original model from timm A_ : List[Any] = timm.create_model(_UpperCAmelCase , pretrained=_UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ : Tuple = timm_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) A_ : Any = create_rename_keys(_UpperCAmelCase , _UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : Union[str, Any] = 'huggingface/label-files' A_ : Dict = 'imagenet-1k-id2label.json' A_ : List[str] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) A_ : str = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} A_ : Any = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": A_ : List[Any] = ViTHybridModel(_UpperCAmelCase ).eval() else: A_ : str = ViTHybridForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # create image processor A_ : Dict = create_transform(**resolve_data_config({} , model=_UpperCAmelCase ) ) A_ : List[str] = transform.transforms A_ : List[str] = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } A_ : Tuple = ViTHybridImageProcessor( do_resize=_UpperCAmelCase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_UpperCAmelCase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=_UpperCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) A_ : Optional[Any] = prepare_img() A_ : Any = transform(_UpperCAmelCase ).unsqueeze(0 ) A_ : Dict = processor(_UpperCAmelCase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase ) # verify logits with torch.no_grad(): A_ : List[Any] = model(_UpperCAmelCase ) A_ : List[str] = outputs.logits print('Predicted class:' , logits.argmax(-1 ).item() ) if base_model: A_ : Union[str, Any] = timm_model.forward_features(_UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_UpperCAmelCase , outputs.pooler_output , atol=1E-3 ) else: A_ : Tuple = timm_model(_UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase , outputs.logits , atol=1E-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(f"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: print(f"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(f"""ybelkada/{vit_name}""" ) processor.push_to_hub(f"""ybelkada/{vit_name}""" ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_r50_s16_384', type=str, help='Name of the hybrid ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) lowerCamelCase_ : List[str] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
286
0
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if isinstance(lowerCamelCase__ , torch.Tensor ): return image elif isinstance(lowerCamelCase__ , PIL.Image.Image ): lowerCamelCase_ = [image] if isinstance(image[0] , PIL.Image.Image ): lowerCamelCase_ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) )[None, :] for i in image] lowerCamelCase_ = np.concatenate(lowerCamelCase__ , axis=0 ) lowerCamelCase_ = np.array(lowerCamelCase__ ).astype(np.floataa ) / 2_55.0 lowerCamelCase_ = image.transpose(0 , 3 , 1 , 2 ) lowerCamelCase_ = 2.0 * image - 1.0 lowerCamelCase_ = torch.from_numpy(lowerCamelCase__ ) elif isinstance(image[0] , torch.Tensor ): lowerCamelCase_ = torch.cat(lowerCamelCase__ , dim=0 ) return image def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=0.99_95 ): if not isinstance(lowerCamelCase__ , np.ndarray ): lowerCamelCase_ = True lowerCamelCase_ = va.device lowerCamelCase_ = va.cpu().numpy() lowerCamelCase_ = va.cpu().numpy() lowerCamelCase_ = np.sum(va * va / (np.linalg.norm(lowerCamelCase__ ) * np.linalg.norm(lowerCamelCase__ )) ) if np.abs(lowerCamelCase__ ) > DOT_THRESHOLD: lowerCamelCase_ = (1 - t) * va + t * va else: lowerCamelCase_ = np.arccos(lowerCamelCase__ ) lowerCamelCase_ = np.sin(lowerCamelCase__ ) lowerCamelCase_ = theta_a * t lowerCamelCase_ = np.sin(lowerCamelCase__ ) lowerCamelCase_ = np.sin(theta_a - theta_t ) / sin_theta_a lowerCamelCase_ = sin_theta_t / sin_theta_a lowerCamelCase_ = sa * va + sa * va if inputs_are_torch: lowerCamelCase_ = torch.from_numpy(lowerCamelCase__ ).to(lowerCamelCase__ ) return va def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = F.normalize(lowerCamelCase__ , dim=-1 ) lowerCamelCase_ = F.normalize(lowerCamelCase__ , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): for param in model.parameters(): lowerCamelCase_ = value class _SCREAMING_SNAKE_CASE ( snake_case_ ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None , lowercase=None , lowercase=None , ) -> Union[str, Any]: super().__init__() self.register_modules( vae=lowercase , text_encoder=lowercase , clip_model=lowercase , tokenizer=lowercase , unet=lowercase , scheduler=lowercase , feature_extractor=lowercase , coca_model=lowercase , coca_tokenizer=lowercase , coca_transform=lowercase , ) lowerCamelCase_ = ( feature_extractor.size if isinstance(feature_extractor.size , lowercase ) else feature_extractor.size["shortest_edge"] ) lowerCamelCase_ = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , lowercase ) set_requires_grad(self.clip_model , lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase = "auto" ) -> Union[str, Any]: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCamelCase_ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: self.enable_attention_slicing(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> int: set_requires_grad(self.vae , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: set_requires_grad(self.vae , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: set_requires_grad(self.unet , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: set_requires_grad(self.unet , lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase ) -> List[Any]: # get the original timestep using init_timestep lowerCamelCase_ = min(int(num_inference_steps * strength ) , lowercase ) lowerCamelCase_ = max(num_inference_steps - init_timestep , 0 ) lowerCamelCase_ = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None ) -> Any: if not isinstance(lowercase , torch.Tensor ): raise ValueError(f'`image` has to be of type `torch.Tensor` but is {type(lowercase )}' ) lowerCamelCase_ = image.to(device=lowercase , dtype=lowercase ) if isinstance(lowercase , lowercase ): lowerCamelCase_ = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(lowercase ) ] lowerCamelCase_ = torch.cat(lowercase , dim=0 ) else: lowerCamelCase_ = self.vae.encode(lowercase ).latent_dist.sample(lowercase ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor lowerCamelCase_ = 0.1_8_2_1_5 * init_latents lowerCamelCase_ = init_latents.repeat_interleave(lowercase , dim=0 ) lowerCamelCase_ = randn_tensor(init_latents.shape , generator=lowercase , device=lowercase , dtype=lowercase ) # get latents lowerCamelCase_ = self.scheduler.add_noise(lowercase , lowercase , lowercase ) lowerCamelCase_ = init_latents return latents def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Tuple: lowerCamelCase_ = self.coca_transform(lowercase ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): lowerCamelCase_ = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) lowerCamelCase_ = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("<end_of_text>" )[0].replace("<start_of_text>" , "" ).rstrip(" .," ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> str: lowerCamelCase_ = self.feature_extractor.preprocess(lowercase ) lowerCamelCase_ = torch.from_numpy(clip_image_input["pixel_values"][0] ).unsqueeze(0 ).to(self.device ).half() lowerCamelCase_ = self.clip_model.get_image_features(lowercase ) lowerCamelCase_ = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=lowercase ) lowerCamelCase_ = image_embeddings_clip.repeat_interleave(lowercase , dim=0 ) return image_embeddings_clip @torch.enable_grad() def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Optional[Any]: lowerCamelCase_ = latents.detach().requires_grad_() lowerCamelCase_ = self.scheduler.scale_model_input(lowercase , lowercase ) # predict the noise residual lowerCamelCase_ = self.unet(lowercase , lowercase , encoder_hidden_states=lowercase ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): lowerCamelCase_ = self.scheduler.alphas_cumprod[timestep] lowerCamelCase_ = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowerCamelCase_ = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 lowerCamelCase_ = torch.sqrt(lowercase ) lowerCamelCase_ = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , lowercase ): lowerCamelCase_ = self.scheduler.sigmas[index] lowerCamelCase_ = latents - sigma * noise_pred else: raise ValueError(f'scheduler type {type(self.scheduler )} not supported' ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor lowerCamelCase_ = 1 / 0.1_8_2_1_5 * sample lowerCamelCase_ = self.vae.decode(lowercase ).sample lowerCamelCase_ = (image / 2 + 0.5).clamp(0 , 1 ) lowerCamelCase_ = transforms.Resize(self.feature_extractor_size )(lowercase ) lowerCamelCase_ = self.normalize(lowercase ).to(latents.dtype ) lowerCamelCase_ = self.clip_model.get_image_features(lowercase ) lowerCamelCase_ = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=lowercase ) lowerCamelCase_ = spherical_dist_loss(lowercase , lowercase ).mean() * clip_guidance_scale lowerCamelCase_ = -torch.autograd.grad(lowercase , lowercase )[0] if isinstance(self.scheduler , lowercase ): lowerCamelCase_ = latents.detach() + grads * (sigma**2) lowerCamelCase_ = noise_pred_original else: lowerCamelCase_ = noise_pred_original - torch.sqrt(lowercase ) * grads return noise_pred, latents @torch.no_grad() def __call__( self , lowercase , lowercase , lowercase = None , lowercase = None , lowercase = 512 , lowercase = 512 , lowercase = 0.6 , lowercase = 50 , lowercase = 7.5 , lowercase = 1 , lowercase = 0.0 , lowercase = 100 , lowercase = None , lowercase = "pil" , lowercase = True , lowercase = 0.8 , lowercase = 0.1 , lowercase = 0.1 , ) -> int: if isinstance(lowercase , lowercase ) and len(lowercase ) != batch_size: raise ValueError(f'You have passed {batch_size} batch_size, but only {len(lowercase )} generators.' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'`height` and `width` have to be divisible by 8 but are {height} and {width}.' ) if isinstance(lowercase , torch.Generator ) and batch_size > 1: lowerCamelCase_ = [generator] + [None] * (batch_size - 1) lowerCamelCase_ = [ ("model", self.coca_model is None), ("tokenizer", self.coca_tokenizer is None), ("transform", self.coca_transform is None), ] lowerCamelCase_ = [x[0] for x in coca_is_none if x[1]] lowerCamelCase_ = ", ".join(lowercase ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(lowercase ): raise ValueError( f'Content prompt is None and CoCa [{coca_is_none_str}] is None.' f'Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.' ) lowerCamelCase_ = self.get_image_description(lowercase ) if style_prompt is None: if len(lowercase ): raise ValueError( f'Style prompt is None and CoCa [{coca_is_none_str}] is None.' f' Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.' ) lowerCamelCase_ = self.get_image_description(lowercase ) # get prompt text embeddings for content and style lowerCamelCase_ = self.tokenizer( lowercase , padding="max_length" , max_length=self.tokenizer.model_max_length , truncation=lowercase , return_tensors="pt" , ) lowerCamelCase_ = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] lowerCamelCase_ = self.tokenizer( lowercase , padding="max_length" , max_length=self.tokenizer.model_max_length , truncation=lowercase , return_tensors="pt" , ) lowerCamelCase_ = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] lowerCamelCase_ = slerp(lowercase , lowercase , lowercase ) # duplicate text embeddings for each generation per prompt lowerCamelCase_ = text_embeddings.repeat_interleave(lowercase , dim=0 ) # set timesteps lowerCamelCase_ = "offset" in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) lowerCamelCase_ = {} if accepts_offset: lowerCamelCase_ = 1 self.scheduler.set_timesteps(lowercase , **lowercase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) lowerCamelCase_ , lowerCamelCase_ = self.get_timesteps(lowercase , lowercase , self.device ) lowerCamelCase_ = timesteps[:1].repeat(lowercase ) # Preprocess image lowerCamelCase_ = preprocess(lowercase , lowercase , lowercase ) lowerCamelCase_ = self.prepare_latents( lowercase , lowercase , lowercase , text_embeddings.dtype , self.device , lowercase ) lowerCamelCase_ = preprocess(lowercase , lowercase , lowercase ) lowerCamelCase_ = self.prepare_latents( lowercase , lowercase , lowercase , text_embeddings.dtype , self.device , lowercase ) lowerCamelCase_ = slerp(lowercase , lowercase , lowercase ) if clip_guidance_scale > 0: lowerCamelCase_ = self.get_clip_image_embeddings(lowercase , lowercase ) lowerCamelCase_ = self.get_clip_image_embeddings(lowercase , lowercase ) lowerCamelCase_ = slerp( lowercase , lowercase , lowercase ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. lowerCamelCase_ = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: lowerCamelCase_ = content_text_input.input_ids.shape[-1] lowerCamelCase_ = self.tokenizer([""] , padding="max_length" , max_length=lowercase , return_tensors="pt" ) lowerCamelCase_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt lowerCamelCase_ = uncond_embeddings.repeat_interleave(lowercase , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowerCamelCase_ = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. lowerCamelCase_ = (batch_size, self.unet.config.in_channels, height // 8, width // 8) lowerCamelCase_ = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps lowerCamelCase_ = torch.randn(lowercase , generator=lowercase , device="cpu" , dtype=lowercase ).to( self.device ) else: lowerCamelCase_ = torch.randn(lowercase , generator=lowercase , device=self.device , dtype=lowercase ) else: if latents.shape != latents_shape: raise ValueError(f'Unexpected latents shape, got {latents.shape}, expected {latents_shape}' ) lowerCamelCase_ = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler lowerCamelCase_ = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCamelCase_ = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCamelCase_ = {} if accepts_eta: lowerCamelCase_ = eta # check if the scheduler accepts generator lowerCamelCase_ = "generator" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: lowerCamelCase_ = generator with self.progress_bar(total=lowercase ): for i, t in enumerate(lowercase ): # expand the latents if we are doing classifier free guidance lowerCamelCase_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase_ = self.scheduler.scale_model_input(lowercase , lowercase ) # predict the noise residual lowerCamelCase_ = self.unet(lowercase , lowercase , encoder_hidden_states=lowercase ).sample # perform classifier free guidance if do_classifier_free_guidance: lowerCamelCase_ , lowerCamelCase_ = noise_pred.chunk(2 ) lowerCamelCase_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: lowerCamelCase_ = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) lowerCamelCase_ , lowerCamelCase_ = self.cond_fn( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase_ = self.scheduler.step(lowercase , lowercase , lowercase , **lowercase ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor lowerCamelCase_ = 1 / 0.1_8_2_1_5 * latents lowerCamelCase_ = self.vae.decode(lowercase ).sample lowerCamelCase_ = (image / 2 + 0.5).clamp(0 , 1 ) lowerCamelCase_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(lowercase ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=lowercase , nsfw_content_detected=lowercase )
19
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" A_ : str = False if num < 0: A_ : Dict = True A_ : Union[str, Any] = -num A_ : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(_UpperCAmelCase ) for e in binary ) return "0b" + "".join(str(_UpperCAmelCase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
286
0
import numpy as np lowercase : List[Any] = [ ["""a""", """b""", """c""", """d""", """e"""], ["""f""", """g""", """h""", """i""", """k"""], ["""l""", """m""", """n""", """o""", """p"""], ["""q""", """r""", """s""", """t""", """u"""], ["""v""", """w""", """x""", """y""", """z"""], ] class __snake_case : def __init__( self ): '''simple docstring''' lowercase : int = np.array(snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase , lowercase : Tuple = np.where(letter == self.SQUARE ) lowercase : List[str] = np.concatenate([indexa + 1, indexa + 1] ) return indexes def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : str = self.SQUARE[indexa - 1, indexa - 1] return letter def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : List[str] = message.lower() lowercase : str = message.replace(""" """ ,"""""" ) lowercase : Tuple = message.replace("""j""" ,"""i""" ) lowercase : List[str] = np.empty((2, len(snake_case )) ) for letter_index in range(len(snake_case ) ): lowercase : List[str] = self.letter_to_numbers(message[letter_index] ) lowercase : List[str] = numbers[0] lowercase : Union[str, Any] = numbers[1] lowercase : str = first_step.reshape(2 * len(snake_case ) ) lowercase : Dict = """""" for numbers_index in range(len(snake_case ) ): lowercase : Any = int(second_step[numbers_index * 2] ) lowercase : Optional[int] = int(second_step[(numbers_index * 2) + 1] ) lowercase : List[str] = self.numbers_to_letter(snake_case ,snake_case ) lowercase : Tuple = encoded_message + letter return encoded_message def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : Optional[Any] = message.lower() message.replace(""" """ ,"""""" ) lowercase : Dict = np.empty(2 * len(snake_case ) ) for letter_index in range(len(snake_case ) ): lowercase : Optional[int] = self.letter_to_numbers(message[letter_index] ) lowercase : List[str] = numbers[0] lowercase : Any = numbers[1] lowercase : Optional[int] = first_step.reshape((2, len(snake_case )) ) lowercase : Tuple = """""" for numbers_index in range(len(snake_case ) ): lowercase : Any = int(second_step[0, numbers_index] ) lowercase : Optional[int] = int(second_step[1, numbers_index] ) lowercase : List[str] = self.numbers_to_letter(snake_case ,snake_case ) lowercase : Any = decoded_message + letter return decoded_message
20
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCamelCase_ : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Tuple = ['MLukeTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys lowerCamelCase_ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
0
from __future__ import annotations from collections.abc import Generator def UpperCamelCase_( ) -> Generator[int, None, None]: _lowercase : dict[int, int] = {} _lowercase : Optional[int] = 2 while True: _lowercase : List[Any] = factor_map.pop(lowerCamelCase_ , lowerCamelCase_ ) if factor: _lowercase : Optional[Any] = factor + prime while x in factor_map: x += factor _lowercase : Optional[int] = factor else: _lowercase : Union[str, Any] = prime yield prime prime += 1 def UpperCamelCase_( lowerCamelCase_ = 1e10 ) -> int: _lowercase : Optional[int] = sieve() _lowercase : int = 1 while True: _lowercase : Any = next(lowerCamelCase_ ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(lowerCamelCase_ ) n += 2 if __name__ == "__main__": print(solution())
21
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers lowerCamelCase_ : List[str] = [int(0.5 * n * (n + 1)) for n in range(1, 1_01)] def UpperCAmelCase__ ( ): """simple docstring""" A_ : Union[str, Any] = os.path.dirname(os.path.realpath(_UpperCAmelCase ) ) A_ : Tuple = os.path.join(_UpperCAmelCase , 'words.txt' ) A_ : List[Any] = '' with open(_UpperCAmelCase ) as f: A_ : int = f.readline() A_ : Optional[Any] = [word.strip('"' ) for word in words.strip('\r\n' ).split(',' )] A_ : Dict = [ word for word in [sum(ord(_UpperCAmelCase ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
286
0
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE :Optional[int] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE :str = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class A_ ( lowerCAmelCase_ ): _lowerCamelCase : Tuple = """encodec""" def __init__( self : Tuple , snake_case_ : Optional[int]=[1.5, 3.0, 6.0, 1_2.0, 2_4.0] , snake_case_ : Union[str, Any]=2_4_0_0_0 , snake_case_ : List[Any]=1 , snake_case_ : Optional[int]=False , snake_case_ : Any=None , snake_case_ : Dict=None , snake_case_ : str=1_2_8 , snake_case_ : Union[str, Any]=3_2 , snake_case_ : Optional[int]=1 , snake_case_ : Optional[int]=[8, 5, 4, 2] , snake_case_ : Dict="weight_norm" , snake_case_ : List[str]=7 , snake_case_ : Optional[Any]=7 , snake_case_ : Dict=3 , snake_case_ : Dict=2 , snake_case_ : Tuple=True , snake_case_ : Tuple="reflect" , snake_case_ : Tuple=2 , snake_case_ : Tuple=2 , snake_case_ : Union[str, Any]=1.0 , snake_case_ : List[str]=1_0_2_4 , snake_case_ : Union[str, Any]=None , snake_case_ : List[Any]=True , **snake_case_ : List[Any] , ): _UpperCAmelCase = target_bandwidths _UpperCAmelCase = sampling_rate _UpperCAmelCase = audio_channels _UpperCAmelCase = normalize _UpperCAmelCase = chunk_length_s _UpperCAmelCase = overlap _UpperCAmelCase = hidden_size _UpperCAmelCase = num_filters _UpperCAmelCase = num_residual_layers _UpperCAmelCase = upsampling_ratios _UpperCAmelCase = norm_type _UpperCAmelCase = kernel_size _UpperCAmelCase = last_kernel_size _UpperCAmelCase = residual_kernel_size _UpperCAmelCase = dilation_growth_rate _UpperCAmelCase = use_causal_conv _UpperCAmelCase = pad_mode _UpperCAmelCase = compress _UpperCAmelCase = num_lstm_layers _UpperCAmelCase = trim_right_ratio _UpperCAmelCase = codebook_size _UpperCAmelCase = codebook_dim if codebook_dim is not None else hidden_size _UpperCAmelCase = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def lowercase ( self : Optional[int] ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def lowercase ( self : Tuple ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def lowercase ( self : Dict ): _UpperCAmelCase = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def lowercase ( self : Union[str, Any] ): return int(1_0_0_0 * self.target_bandwidths[-1] // (self.frame_rate * 1_0) )
22
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ : List[str] = {'configuration_xlnet': ['XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : str = ['XLNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[str] = ['XLNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ 'XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLNetForMultipleChoice', 'XLNetForQuestionAnswering', 'XLNetForQuestionAnsweringSimple', 'XLNetForSequenceClassification', 'XLNetForTokenClassification', 'XLNetLMHeadModel', 'XLNetModel', 'XLNetPreTrainedModel', 'load_tf_weights_in_xlnet', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Union[str, Any] = [ 'TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLNetForMultipleChoice', 'TFXLNetForQuestionAnsweringSimple', 'TFXLNetForSequenceClassification', 'TFXLNetForTokenClassification', 'TFXLNetLMHeadModel', 'TFXLNetMainLayer', 'TFXLNetModel', 'TFXLNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
286
0
'''simple docstring''' import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase__: str = logging.get_logger(__name__) UpperCamelCase__: List[Any] = { "facebook/detr-resnet-50": "https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json", # See all DETR models at https://huggingface.co/models?filter=detr } class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """detr""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self : Tuple , __snake_case : Any=True , __snake_case : int=None , __snake_case : Dict=3 , __snake_case : Optional[Any]=100 , __snake_case : str=6 , __snake_case : Tuple=2048 , __snake_case : int=8 , __snake_case : List[Any]=6 , __snake_case : Optional[int]=2048 , __snake_case : Tuple=8 , __snake_case : Tuple=0.0 , __snake_case : Union[str, Any]=0.0 , __snake_case : str=True , __snake_case : Tuple="relu" , __snake_case : Optional[Any]=256 , __snake_case : Optional[Any]=0.1 , __snake_case : Dict=0.0 , __snake_case : Any=0.0 , __snake_case : Union[str, Any]=0.02 , __snake_case : Tuple=1.0 , __snake_case : Optional[int]=False , __snake_case : Union[str, Any]="sine" , __snake_case : Optional[Any]="resnet50" , __snake_case : str=True , __snake_case : List[Any]=False , __snake_case : Tuple=1 , __snake_case : Union[str, Any]=5 , __snake_case : Optional[int]=2 , __snake_case : int=1 , __snake_case : Optional[int]=1 , __snake_case : Union[str, Any]=5 , __snake_case : Any=2 , __snake_case : Optional[int]=0.1 , **__snake_case : Any , ) -> Any: if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) UpperCAmelCase : Optional[Any] = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(__snake_case , __snake_case ): UpperCAmelCase : Union[str, Any] = backbone_config.get('''model_type''' ) UpperCAmelCase : List[Any] = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase : int = config_class.from_dict(__snake_case ) # set timm attributes to None UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[Any] = None, None, None UpperCAmelCase : Optional[Any] = use_timm_backbone UpperCAmelCase : Union[str, Any] = backbone_config UpperCAmelCase : List[str] = num_channels UpperCAmelCase : Optional[Any] = num_queries UpperCAmelCase : str = d_model UpperCAmelCase : List[Any] = encoder_ffn_dim UpperCAmelCase : Tuple = encoder_layers UpperCAmelCase : str = encoder_attention_heads UpperCAmelCase : List[Any] = decoder_ffn_dim UpperCAmelCase : List[Any] = decoder_layers UpperCAmelCase : List[str] = decoder_attention_heads UpperCAmelCase : List[str] = dropout UpperCAmelCase : Union[str, Any] = attention_dropout UpperCAmelCase : int = activation_dropout UpperCAmelCase : Union[str, Any] = activation_function UpperCAmelCase : Any = init_std UpperCAmelCase : List[str] = init_xavier_std UpperCAmelCase : Dict = encoder_layerdrop UpperCAmelCase : Optional[int] = decoder_layerdrop UpperCAmelCase : Optional[int] = encoder_layers UpperCAmelCase : Tuple = auxiliary_loss UpperCAmelCase : Union[str, Any] = position_embedding_type UpperCAmelCase : Union[str, Any] = backbone UpperCAmelCase : Optional[int] = use_pretrained_backbone UpperCAmelCase : Optional[Any] = dilation # Hungarian matcher UpperCAmelCase : Optional[int] = class_cost UpperCAmelCase : Optional[int] = bbox_cost UpperCAmelCase : Tuple = giou_cost # Loss coefficients UpperCAmelCase : List[str] = mask_loss_coefficient UpperCAmelCase : Union[str, Any] = dice_loss_coefficient UpperCAmelCase : Optional[Any] = bbox_loss_coefficient UpperCAmelCase : Tuple = giou_loss_coefficient UpperCAmelCase : Dict = eos_coefficient super().__init__(is_encoder_decoder=__snake_case , **__snake_case ) @property def A ( self : Optional[Any] ) -> int: return self.encoder_attention_heads @property def A ( self : int ) -> int: return self.d_model @classmethod def A ( cls : List[Any] , __snake_case : PretrainedConfig , **__snake_case : str ) -> str: return cls(backbone_config=__snake_case , **__snake_case ) def A ( self : Union[str, Any] ) -> Dict[str, any]: UpperCAmelCase : List[str] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase : Any = self.backbone_config.to_dict() UpperCAmelCase : int = self.__class__.model_type return output class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = version.parse("""1.11""" ) @property def A ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def A ( self : List[str] ) -> float: return 1E-5 @property def A ( self : str ) -> int: return 12
23
"""simple docstring""" import torch from diffusers import DiffusionPipeline class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" super().__init__() self.register_modules(unet=snake_case_ , scheduler=snake_case_ ) def __call__( self ): """simple docstring""" A_ : Optional[Any] = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) A_ : List[str] = 1 A_ : List[str] = self.unet(snake_case_ , snake_case_ ).sample A_ : Optional[int] = self.scheduler.step(snake_case_ , snake_case_ , snake_case_ ).prev_sample A_ : List[Any] = scheduler_output - scheduler_output + torch.ones_like(snake_case_ ) return result
286
0
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def lowerCamelCase__ ( snake_case_ : List[str] ) -> List[str]: return EnvironmentCommand() def lowerCamelCase__ ( snake_case_ : List[str] ) -> Union[str, Any]: return EnvironmentCommand(args.accelerate_config_file ) class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): @staticmethod def a (a__ : ArgumentParser ): """simple docstring""" __snake_case = parser.add_parser('''env''' ) download_parser.set_defaults(func=a__ ) download_parser.add_argument( '''--accelerate-config_file''' , default=a__ , help='''The accelerate config file to use for the default values in the launching script.''' , ) download_parser.set_defaults(func=a__ ) def __init__(self : Union[str, Any] , a__ : str , *a__ : int ): """simple docstring""" __snake_case = accelerate_config_file def a (self : Union[str, Any] ): """simple docstring""" __snake_case = '''not installed''' if is_safetensors_available(): import safetensors __snake_case = safetensors.__version__ elif importlib.util.find_spec('''safetensors''' ) is not None: import safetensors __snake_case = f"""{safetensors.__version__} but is ignored because of PyTorch version too old.""" __snake_case = '''not installed''' __snake_case = __snake_case = '''not found''' if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file __snake_case = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(a__ ): __snake_case = load_config_from_file(self._accelerate_config_file ).to_dict() __snake_case = ( '''\n'''.join([f"""\t- {prop}: {val}""" for prop, val in accelerate_config.items()] ) if isinstance(a__ , a__ ) else f"""\t{accelerate_config}""" ) __snake_case = '''not installed''' __snake_case = '''NA''' if is_torch_available(): import torch __snake_case = torch.__version__ __snake_case = torch.cuda.is_available() __snake_case = '''not installed''' __snake_case = '''NA''' if is_tf_available(): import tensorflow as tf __snake_case = tf.__version__ try: # deprecated in v2.1 __snake_case = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool __snake_case = bool(tf.config.list_physical_devices('''GPU''' ) ) __snake_case = '''not installed''' __snake_case = '''not installed''' __snake_case = '''not installed''' __snake_case = '''NA''' if is_flax_available(): import flax import jax import jaxlib __snake_case = flax.__version__ __snake_case = jax.__version__ __snake_case = jaxlib.__version__ __snake_case = jax.lib.xla_bridge.get_backend().platform __snake_case = { '''`transformers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Huggingface_hub version''': huggingface_hub.__version__, '''Safetensors version''': f"""{safetensors_version}""", '''Accelerate version''': f"""{accelerate_version}""", '''Accelerate config''': f"""{accelerate_config_str}""", '''PyTorch version (GPU?)''': f"""{pt_version} ({pt_cuda_available})""", '''Tensorflow version (GPU?)''': f"""{tf_version} ({tf_cuda_available})""", '''Flax version (CPU?/GPU?/TPU?)''': f"""{flax_version} ({jax_backend})""", '''Jax version''': f"""{jax_version}""", '''JaxLib version''': f"""{jaxlib_version}""", '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(a__ ) ) return info @staticmethod def a (a__ : Dict ): """simple docstring""" return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
24
"""simple docstring""" from heapq import heappop, heappush import numpy as np def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" A_ , A_ : List[str] = grid.shape A_ : Optional[int] = [-1, 1, 0, 0] A_ : str = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] A_ , A_ : List[Any] = [(0, source)], set() A_ : Optional[Any] = np.full((rows, cols) , np.inf ) A_ : int = 0 A_ : Optional[int] = np.empty((rows, cols) , dtype=_UpperCAmelCase ) A_ : Optional[int] = None while queue: ((A_) , (A_)) : str = heappop(_UpperCAmelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: A_ : int = [] while (x, y) != source: path.append((x, y) ) A_ , A_ : List[Any] = predecessors[x, y] path.append(_UpperCAmelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(_UpperCAmelCase ) ): A_ , A_ : Tuple = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: A_ : Union[str, Any] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(_UpperCAmelCase , (dist + 1, (nx, ny)) ) A_ : Optional[Any] = dist + 1 A_ : Optional[Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
286
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) UpperCAmelCase__ : Optional[Any] = {'configuration_beit': ['BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BeitConfig', 'BeitOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : Optional[Any] = ['BeitFeatureExtractor'] UpperCAmelCase__ : Any = ['BeitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : List[Any] = [ 'BEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BeitForImageClassification', 'BeitForMaskedImageModeling', 'BeitForSemanticSegmentation', 'BeitModel', 'BeitPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : Tuple = [ 'FlaxBeitForImageClassification', 'FlaxBeitForMaskedImageModeling', 'FlaxBeitModel', 'FlaxBeitPreTrainedModel', ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys UpperCAmelCase__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
25
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase_ : Optional[Any] = { 'huggingface/informer-tourism-monthly': ( 'https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json' ), # See all Informer models at https://huggingface.co/models?filter=informer } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """informer""" lowercase_ : str = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self , snake_case_ = None , snake_case_ = None , snake_case_ = "student_t" , snake_case_ = "nll" , snake_case_ = 1 , snake_case_ = None , snake_case_ = "mean" , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = 6_4 , snake_case_ = 3_2 , snake_case_ = 3_2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = True , snake_case_ = "gelu" , snake_case_ = 0.05 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 1_0_0 , snake_case_ = 0.02 , snake_case_=True , snake_case_ = "prob" , snake_case_ = 5 , snake_case_ = True , **snake_case_ , ): """simple docstring""" A_ : str = prediction_length A_ : List[Any] = context_length or prediction_length A_ : str = distribution_output A_ : Dict = loss A_ : Any = input_size A_ : Union[str, Any] = num_time_features A_ : Optional[Any] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] A_ : List[Any] = scaling A_ : Tuple = num_dynamic_real_features A_ : Any = num_static_real_features A_ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) A_ : Optional[int] = cardinality else: A_ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(snake_case_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) A_ : Any = embedding_dimension else: A_ : Optional[Any] = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] A_ : int = num_parallel_samples # Transformer architecture configuration A_ : str = input_size * len(self.lags_sequence ) + self._number_of_features A_ : List[Any] = d_model A_ : Dict = encoder_attention_heads A_ : Dict = decoder_attention_heads A_ : List[Any] = encoder_ffn_dim A_ : Union[str, Any] = decoder_ffn_dim A_ : int = encoder_layers A_ : Any = decoder_layers A_ : List[Any] = dropout A_ : str = attention_dropout A_ : Tuple = activation_dropout A_ : List[str] = encoder_layerdrop A_ : List[str] = decoder_layerdrop A_ : str = activation_function A_ : Optional[int] = init_std A_ : List[Any] = use_cache # Informer A_ : Tuple = attention_type A_ : List[Any] = sampling_factor A_ : Optional[int] = distil super().__init__(is_encoder_decoder=snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
286
0
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowercase ( UpperCamelCase__ ): @slow @require_torch def a__ ( self ) -> Dict: _A : int = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) _A : Tuple = BertTokenizer.from_pretrained("""bert-base-uncased""" ) _A : Dict = bertabert.config.encoder.vocab_size _A : List[str] = tokenizer.sep_token_id _A : Dict = tokenizer.cls_token_id _A : List[Any] = 128 _A : Tuple = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) _A : List[str] = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) _A : Union[str, Any] = train_dataset.select(range(32 ) ) _A : str = val_dataset.select(range(16 ) ) _A : Dict = 4 def _map_to_encoder_decoder_inputs(_a ): # Tokenizer will automatically set [BOS] <text> [EOS] _A : List[Any] = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=_a , max_length=512 ) _A : int = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=_a , max_length=128 ) _A : Union[str, Any] = inputs.input_ids _A : List[Any] = inputs.attention_mask _A : Dict = outputs.input_ids _A : str = outputs.input_ids.copy() _A : Union[str, Any] = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] _A : Dict = outputs.attention_mask assert all(len(_a ) == 512 for x in inputs.input_ids ) assert all(len(_a ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(_a ): _A : int = pred.label_ids _A : str = pred.predictions # all unnecessary tokens are removed _A : Dict = tokenizer.batch_decode(_a , skip_special_tokens=_a ) _A : Optional[Any] = tokenizer.batch_decode(_a , skip_special_tokens=_a ) _A : Tuple = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_a ) )] ) / len(_a ) return {"accuracy": accuracy} # map train dataset _A : Dict = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_a , batch_size=_a , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset _A : Optional[int] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_a , batch_size=_a , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) _A : Optional[int] = self.get_auto_remove_tmp_dir() _A : Any = SeqaSeqTrainingArguments( output_dir=_a , per_device_train_batch_size=_a , per_device_eval_batch_size=_a , predict_with_generate=_a , evaluation_strategy="""steps""" , do_train=_a , do_eval=_a , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer _A : Optional[Any] = SeqaSeqTrainer( model=_a , args=_a , compute_metrics=_compute_metrics , train_dataset=_a , eval_dataset=_a , tokenizer=_a , ) # start training trainer.train()
26
"""simple docstring""" import os def UpperCAmelCase__ ( ): """simple docstring""" A_ : Any = os.path.join(os.path.dirname(_UpperCAmelCase ) , 'num.txt' ) with open(_UpperCAmelCase ) as file_hand: return str(sum(int(_UpperCAmelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
286
0
'''simple docstring''' from __future__ import annotations def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : list[str] | None = None , _SCREAMING_SNAKE_CASE : dict[str, float] | None = None , _SCREAMING_SNAKE_CASE : bool = False , ): __a : str = cipher_alphabet or [chr(_SCREAMING_SNAKE_CASE ) for i in range(97 , 123 )] # If the argument is None or the user provided an empty dictionary if not frequencies_dict: # Frequencies of letters in the english language (how much they show up) __a : int = { 'a': 0.0_8_4_9_7, 'b': 0.0_1_4_9_2, 'c': 0.0_2_2_0_2, 'd': 0.0_4_2_5_3, 'e': 0.1_1_1_6_2, 'f': 0.0_2_2_2_8, 'g': 0.0_2_0_1_5, 'h': 0.0_6_0_9_4, 'i': 0.0_7_5_4_6, 'j': 0.0_0_1_5_3, 'k': 0.0_1_2_9_2, 'l': 0.0_4_0_2_5, 'm': 0.0_2_4_0_6, 'n': 0.0_6_7_4_9, 'o': 0.0_7_5_0_7, 'p': 0.0_1_9_2_9, 'q': 0.0_0_0_9_5, 'r': 0.0_7_5_8_7, 's': 0.0_6_3_2_7, 't': 0.0_9_3_5_6, 'u': 0.0_2_7_5_8, 'v': 0.0_0_9_7_8, 'w': 0.0_2_5_6_0, 'x': 0.0_0_1_5_0, 'y': 0.0_1_9_9_4, 'z': 0.0_0_0_7_7, } else: # Custom frequencies dictionary __a : int = frequencies_dict if not case_sensitive: __a : Dict = ciphertext.lower() # Chi squared statistic values __a : dict[int, tuple[float, str]] = {} # cycle through all of the shifts for shift in range(len(_SCREAMING_SNAKE_CASE ) ): __a : Optional[Any] = '' # decrypt the message with the shift for letter in ciphertext: try: # Try to index the letter in the alphabet __a : Any = (alphabet_letters.index(letter.lower() ) - shift) % len( _SCREAMING_SNAKE_CASE ) decrypted_with_shift += ( alphabet_letters[new_key].upper() if case_sensitive and letter.isupper() else alphabet_letters[new_key] ) except ValueError: # Append the character if it isn't in the alphabet decrypted_with_shift += letter __a : Union[str, Any] = 0.0 # Loop through each letter in the decoded message with the shift for letter in decrypted_with_shift: if case_sensitive: __a : Any = letter.lower() if letter in frequencies: # Get the amount of times the letter occurs in the message __a : Optional[int] = decrypted_with_shift.lower().count(_SCREAMING_SNAKE_CASE ) # Get the excepcted amount of times the letter should appear based # on letter frequencies __a : Optional[Any] = frequencies[letter] * occurrences # Complete the chi squared statistic formula __a : Optional[int] = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value else: if letter.lower() in frequencies: # Get the amount of times the letter occurs in the message __a : Optional[Any] = decrypted_with_shift.count(_SCREAMING_SNAKE_CASE ) # Get the excepcted amount of times the letter should appear based # on letter frequencies __a : Optional[int] = frequencies[letter] * occurrences # Complete the chi squared statistic formula __a : Tuple = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value # Add the data to the chi_squared_statistic_values dictionary __a : Dict = ( chi_squared_statistic, decrypted_with_shift, ) # Get the most likely cipher by finding the cipher with the smallest chi squared # statistic def chi_squared_statistic_values_sorting_key(_SCREAMING_SNAKE_CASE : int ) -> tuple[float, str]: return chi_squared_statistic_values[key] __a : int = min( _SCREAMING_SNAKE_CASE , key=_SCREAMING_SNAKE_CASE , ) # Get all the data from the most likely cipher (key, decoded message) ( ( __a ) , ( __a ) , ) : Optional[int] = chi_squared_statistic_values[most_likely_cipher] # Return the data on the most likely shift return ( most_likely_cipher, most_likely_cipher_chi_squared_value, decoded_most_likely_cipher, )
27
"""simple docstring""" import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger lowerCamelCase_ : Dict = get_logger(__name__) lowerCamelCase_ : List[str] = r'\n Args:\n input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam\n search or log softmax for each vocabulary token when using beam search\n kwargs (`Dict[str, Any]`, *optional*):\n Additional logits processor specific kwargs.\n\n Return:\n `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores.\n\n' class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase : '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ ): """simple docstring""" raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @add_start_docstrings(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ): """simple docstring""" for processor in self: A_ : Tuple = inspect.signature(processor.__call__ ).parameters if len(snake_case_ ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( F"""Make sure that all the required parameters: {list(function_args.keys() )} for """ F"""{processor.__class__} are passed to the logits processor.""" ) A_ : Tuple = processor(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) else: A_ : Optional[Any] = processor(snake_case_ , snake_case_ , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or not (temperature > 0): raise ValueError(F"""`temperature` has to be a strictly positive float, but is {temperature}""" ) A_ : Optional[int] = temperature def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = scores / self.temperature return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or (top_p < 0 or top_p > 1.0): raise ValueError(F"""`top_p` has to be a float > 0 and < 1, but is {top_p}""" ) if not isinstance(snake_case_ , snake_case_ ) or (min_tokens_to_keep < 1): raise ValueError(F"""`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}""" ) A_ : str = top_p A_ : Union[str, Any] = filter_value A_ : int = min_tokens_to_keep def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : Tuple = lax.top_k(snake_case_ , scores.shape[-1] ) A_ : List[Any] = jnp.full_like(snake_case_ , self.filter_value ) A_ : List[str] = jax.nn.softmax(snake_case_ , axis=-1 ).cumsum(axis=-1 ) A_ : Optional[int] = cumulative_probs < self.top_p # include the token that is higher than top_p as well A_ : Union[str, Any] = jnp.roll(snake_case_ , 1 ) score_mask |= score_mask.at[:, 0].set(snake_case_ ) # min tokens to keep A_ : int = score_mask.at[:, : self.min_tokens_to_keep].set(snake_case_ ) A_ : Optional[Any] = jnp.where(snake_case_ , snake_case_ , snake_case_ ) A_ : List[Any] = jax.lax.sort_key_val(snake_case_ , snake_case_ )[-1] return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ = -float('Inf' ) , snake_case_ = 1 ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or top_k <= 0: raise ValueError(F"""`top_k` has to be a strictly positive integer, but is {top_k}""" ) A_ : str = max(snake_case_ , snake_case_ ) A_ : Union[str, Any] = filter_value def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ , A_ : int = scores.shape A_ : Tuple = jnp.full(batch_size * vocab_size , self.filter_value ) A_ : Union[str, Any] = min(self.top_k , scores.shape[-1] ) # Safety check A_ , A_ : Dict = lax.top_k(snake_case_ , snake_case_ ) A_ : Optional[int] = jnp.broadcast_to((jnp.arange(snake_case_ ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() A_ : int = topk_scores.flatten() A_ : Any = topk_indices.flatten() + shift A_ : List[str] = next_scores_flat.at[topk_indices_flat].set(snake_case_ ) A_ : Union[str, Any] = next_scores_flat.reshape(snake_case_ , snake_case_ ) return next_scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = bos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Union[str, Any] = 1 - jnp.bool_(cur_len - 1 ) A_ : str = jnp.where(snake_case_ , new_scores.at[:, self.bos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = max_length A_ : Optional[int] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = jnp.full(scores.shape , -float('inf' ) ) A_ : Dict = 1 - jnp.bool_(cur_len - self.max_length + 1 ) A_ : Dict = jnp.where(snake_case_ , new_scores.at[:, self.eos_token_id].set(0 ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" if not isinstance(snake_case_ , snake_case_ ) or min_length < 0: raise ValueError(F"""`min_length` has to be a positive integer, but is {min_length}""" ) if not isinstance(snake_case_ , snake_case_ ) or eos_token_id < 0: raise ValueError(F"""`eos_token_id` has to be a positive integer, but is {eos_token_id}""" ) A_ : Any = min_length A_ : List[Any] = eos_token_id def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : int = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) A_ : Optional[Any] = jnp.where(snake_case_ , scores.at[:, self.eos_token_id].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) A_ : Tuple = begin_index def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Dict = 1 - jnp.bool_(cur_len - self.begin_index ) A_ : int = jnp.where(snake_case_ , scores.at[:, self.begin_suppress_tokens].set(-float('inf' ) ) , snake_case_ ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : List[Any] = list(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Optional[Any] = scores.at[..., self.suppress_tokens].set(-float('inf' ) ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ ): """simple docstring""" A_ : Any = dict(snake_case_ ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. A_ : Tuple = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: A_ : Tuple = force_token_array.at[index].set(snake_case_ ) A_ : Any = jnp.intaa(snake_case_ ) def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" def _force_token(snake_case_ ): A_ : List[Any] = scores.shape[0] A_ : Any = self.force_token_array[generation_idx] A_ : Tuple = jnp.ones_like(snake_case_ , dtype=scores.dtype ) * -float('inf' ) A_ : List[Any] = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) A_ : int = lax.dynamic_update_slice(snake_case_ , snake_case_ , (0, current_token) ) return new_scores A_ : int = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(snake_case_ ) , lambda: scores , ) , ) return scores class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : Tuple = generate_config.eos_token_id A_ : Optional[int] = generate_config.no_timestamps_token_id A_ : List[str] = generate_config.no_timestamps_token_id + 1 A_ : Any = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(snake_case_ , 'max_initial_timestamp_index' ): A_ : List[Any] = generate_config.max_initial_timestamp_index else: A_ : Any = model_config.vocab_size if self.max_initial_timestamp_index is None: A_ : Optional[Any] = model_config.vocab_size def __call__( self , snake_case_ , snake_case_ , snake_case_ ): """simple docstring""" A_ : List[str] = scores.at[:, self.no_timestamps_token_id].set(-float('inf' ) ) def handle_pairs(snake_case_ , snake_case_ ): A_ : Any = jnp.where((cur_len - self.begin_index) >= 1 , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , snake_case_ , ) A_ : Tuple = jnp.where((cur_len - self.begin_index) < 2 , snake_case_ , snake_case_ ) A_ : Any = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , snake_case_ , snake_case_ , ) return jnp.where( snake_case_ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float('inf' ) ) , scores_k.at[: self.eos_token_id].set(-float('inf' ) ) , ) , snake_case_ , ) A_ : Tuple = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) A_ : Optional[Any] = jnp.where(cur_len == self.begin_index , snake_case_ , snake_case_ ) A_ : Tuple = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , snake_case_ , ) A_ : int = self.timestamp_begin + self.max_initial_timestamp_index A_ : List[Any] = jnp.where( snake_case_ , scores.at[:, last_allowed + 1 :].set(-float('inf' ) ) , snake_case_ , ) # if sum of probability over timestamps is above any other token, sample timestamp A_ : Any = jax.nn.log_softmax(snake_case_ , axis=-1 ) def handle_cumulative_probs(snake_case_ , snake_case_ ): A_ : Dict = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) A_ : Optional[Any] = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float('inf' ) ) , snake_case_ , ) A_ : Union[str, Any] = jax.vmap(snake_case_ )(snake_case_ , snake_case_ ) return scores
286
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor _lowerCamelCase : str = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __init__( self : List[Any] , *UpperCamelCase__ : Optional[int] , **UpperCamelCase__ : List[str] ): """simple docstring""" warnings.warn( 'The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use GLPNImageProcessor instead.' , UpperCamelCase__ , ) super().__init__(*UpperCamelCase__ , **UpperCamelCase__ )
28
"""simple docstring""" import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = R'\w+[.]\d+' A_ : int = re.findall(_UpperCAmelCase , _UpperCAmelCase ) for pat in pats: A_ : Optional[int] = key.replace(_UpperCAmelCase , '_'.join(pat.split('.' ) ) ) return key def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): A_ : Union[str, Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: A_ : List[str] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: A_ : Optional[Any] = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer A_ : int = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: A_ : str = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer A_ : Optional[Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": A_ : Optional[Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight A_ : Tuple = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias A_ : Optional[int] = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=42 ): """simple docstring""" A_ : int = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params A_ : Union[str, Any] = flax_model.init_weights(PRNGKey(_UpperCAmelCase ) ) A_ : Optional[Any] = flatten_dict(_UpperCAmelCase ) A_ : Tuple = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): A_ : Any = rename_key(_UpperCAmelCase ) A_ : List[str] = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters A_ , A_ : Union[str, Any] = rename_key_and_reshape_tensor(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown A_ : str = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase )
286
0
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } __UpperCAmelCase = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def lowercase__ ( __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : Optional[Any] ): '''simple docstring''' for attribute in key.split('.' ): UpperCAmelCase_ : List[Any] = getattr(__snake_case , __snake_case ) if weight_type is not None: UpperCAmelCase_ : Any = getattr(__snake_case , __snake_case ).shape else: UpperCAmelCase_ : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": UpperCAmelCase_ : List[str] = value elif weight_type == "weight_g": UpperCAmelCase_ : Optional[int] = value elif weight_type == "weight_v": UpperCAmelCase_ : str = value elif weight_type == "bias": UpperCAmelCase_ : Tuple = value else: UpperCAmelCase_ : Optional[Any] = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def lowercase__ ( __snake_case : Optional[int] , __snake_case : Any ): '''simple docstring''' UpperCAmelCase_ : List[Any] = [] UpperCAmelCase_ : List[Any] = fairseq_model.state_dict() UpperCAmelCase_ : int = hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase_ : Any = False if "conv_layers" in name: load_conv_layer( __snake_case , __snake_case , __snake_case , __snake_case , hf_model.config.feat_extract_norm == 'group' , ) UpperCAmelCase_ : int = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: UpperCAmelCase_ : Optional[Any] = True if "*" in mapped_key: UpperCAmelCase_ : int = name.split(__snake_case )[0].split('.' )[-2] UpperCAmelCase_ : Dict = mapped_key.replace('*' , __snake_case ) if "weight_g" in name: UpperCAmelCase_ : Optional[Any] = 'weight_g' elif "weight_v" in name: UpperCAmelCase_ : List[Any] = 'weight_v' elif "bias" in name and "relative_attention_bias" not in name: UpperCAmelCase_ : int = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase_ : Optional[int] = 'weight' else: UpperCAmelCase_ : Union[str, Any] = None set_recursively(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) continue if not is_used: unused_weights.append(__snake_case ) logger.warning(F"Unused weights: {unused_weights}" ) def lowercase__ ( __snake_case : Any , __snake_case : Tuple , __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : List[Any] ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = full_name.split('conv_layers.' )[-1] UpperCAmelCase_ : List[Any] = name.split('.' ) UpperCAmelCase_ : Dict = int(items[0] ) UpperCAmelCase_ : Optional[int] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) UpperCAmelCase_ : int = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) UpperCAmelCase_ : str = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) UpperCAmelCase_ : Optional[Any] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) UpperCAmelCase_ : List[Any] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(__snake_case ) @torch.no_grad() def lowercase__ ( __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Tuple=None ): '''simple docstring''' UpperCAmelCase_ : List[str] = torch.load(__snake_case ) UpperCAmelCase_ : List[Any] = WavLMConfigOrig(checkpoint['cfg'] ) UpperCAmelCase_ : Union[str, Any] = WavLMOrig(__snake_case ) model.load_state_dict(checkpoint['model'] ) model.eval() if config_path is not None: UpperCAmelCase_ : Optional[Any] = WavLMConfig.from_pretrained(__snake_case ) else: UpperCAmelCase_ : Union[str, Any] = WavLMConfig() UpperCAmelCase_ : Tuple = WavLMModel(__snake_case ) recursively_load_weights(__snake_case , __snake_case ) hf_wavlm.save_pretrained(__snake_case ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') __UpperCAmelCase = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
29
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : List[str] = CustomTokenizer pass
286
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __a = { 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
30
"""simple docstring""" from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch lowerCamelCase_ : str = logging.get_logger(__name__) @add_end_docstrings( UpperCAmelCase__ , r""" top_k (`int`, defaults to 5): The number of predictions to return. targets (`str` or `List[str]`, *optional*): When passed, the model will limit the scores to the passed targets instead of looking up in the whole vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting token will be used (with a warning, and that might be slower). """ , ) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if self.framework == "tf": A_ : str = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": A_ : List[str] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ) else: raise ValueError('Unsupported framework' ) return masked_index def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = self.get_masked_index(snake_case_ ) A_ : str = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['input_ids'][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_=None , **snake_case_ ): """simple docstring""" if return_tensors is None: A_ : Any = self.framework A_ : Dict = self.tokenizer(snake_case_ , return_tensors=snake_case_ ) self.ensure_exactly_one_mask_token(snake_case_ ) return model_inputs def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Dict = self.model(**snake_case_ ) A_ : Optional[int] = model_inputs['input_ids'] return model_outputs def lowerCamelCase_ ( self , snake_case_ , snake_case_=5 , snake_case_=None ): """simple docstring""" if target_ids is not None and target_ids.shape[0] < top_k: A_ : str = target_ids.shape[0] A_ : Optional[Any] = model_outputs['input_ids'][0] A_ : List[Any] = model_outputs['logits'] if self.framework == "tf": A_ : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] A_ : Union[str, Any] = outputs.numpy() A_ : Optional[int] = outputs[0, masked_index, :] A_ : Optional[Any] = stable_softmax(snake_case_ , axis=-1 ) if target_ids is not None: A_ : Union[str, Any] = tf.gather_nd(tf.squeeze(snake_case_ , 0 ) , target_ids.reshape(-1 , 1 ) ) A_ : Optional[int] = tf.expand_dims(snake_case_ , 0 ) A_ : Any = tf.math.top_k(snake_case_ , k=snake_case_ ) A_ , A_ : str = topk.values.numpy(), topk.indices.numpy() else: A_ : int = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=snake_case_ ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample A_ : Tuple = outputs[0, masked_index, :] A_ : List[str] = logits.softmax(dim=-1 ) if target_ids is not None: A_ : str = probs[..., target_ids] A_ , A_ : List[str] = probs.topk(snake_case_ ) A_ : List[Any] = [] A_ : int = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): A_ : str = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place A_ : Union[str, Any] = input_ids.numpy().copy() if target_ids is not None: A_ : str = target_ids[p].tolist() A_ : Union[str, Any] = p # Filter padding out: A_ : Any = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back A_ : Any = self.tokenizer.decode(snake_case_ , skip_special_tokens=snake_case_ ) A_ : Any = {'score': v, 'token': p, 'token_str': self.tokenizer.decode([p] ), 'sequence': sequence} row.append(snake_case_ ) result.append(snake_case_ ) if single_mask: return result[0] return result def lowerCamelCase_ ( self , snake_case_ , snake_case_=None ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : List[str] = [targets] try: A_ : Optional[int] = self.tokenizer.get_vocab() except Exception: A_ : int = {} A_ : Tuple = [] for target in targets: A_ : int = vocab.get(snake_case_ , snake_case_ ) if id_ is None: A_ : Tuple = self.tokenizer( snake_case_ , add_special_tokens=snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , max_length=1 , truncation=snake_case_ , )['input_ids'] if len(snake_case_ ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ 'We cannot replace it with anything meaningful, ignoring it' ) continue A_ : str = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) A_ : Tuple = list(set(snake_case_ ) ) if len(snake_case_ ) == 0: raise ValueError('At least one target must be provided when passed.' ) A_ : Optional[Any] = np.array(snake_case_ ) return target_ids def lowerCamelCase_ ( self , snake_case_=None , snake_case_=None ): """simple docstring""" A_ : List[str] = {} if targets is not None: A_ : Any = self.get_target_ids(snake_case_ , snake_case_ ) A_ : Optional[Any] = target_ids if top_k is not None: A_ : int = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( 'fill-mask' , self.model.base_model_prefix , 'The tokenizer does not define a `mask_token`.' ) return {}, {}, postprocess_params def __call__( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : List[str] = super().__call__(snake_case_ , **snake_case_ ) if isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1: return outputs[0] return outputs
286
0
'''simple docstring''' from __future__ import annotations __SCREAMING_SNAKE_CASE : Optional[int] = list[list[int]] # assigning initial values to the grid __SCREAMING_SNAKE_CASE : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution __SCREAMING_SNAKE_CASE : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def UpperCamelCase_ ( _UpperCAmelCase : Matrix , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> bool: """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> tuple[int, int] | None: """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> Matrix | None: """simple docstring""" if location := find_empty_location(_UpperCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : int = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): _UpperCAmelCase : Dict = digit if sudoku(_UpperCAmelCase ) is not None: return grid _UpperCAmelCase : Tuple = 0 return None def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> None: """simple docstring""" for row in grid: for cell in row: print(_UpperCAmelCase , end=" " ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("""\nExample grid:\n""" + """=""" * 20) print_solution(example_grid) print("""\nExample grid solution:""") __SCREAMING_SNAKE_CASE : Tuple = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("""Cannot find a solution.""")
31
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" super().__init__(*snake_case_ , **snake_case_ ) A_ : Tuple = {} def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , **snake_case_ ): """simple docstring""" A_ : str = super().add_tokens(snake_case_ , *snake_case_ , **snake_case_ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.' ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=1 , **snake_case_ ): """simple docstring""" A_ : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) else: A_ : Tuple = [] for i in range(snake_case_ ): A_ : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(snake_case_ , *snake_case_ , **snake_case_ ) output.append(snake_case_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) A_ : Any = output def lowerCamelCase_ ( self , snake_case_ , snake_case_=False , snake_case_=1.0 ): """simple docstring""" if isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = [] for i in range(len(snake_case_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=snake_case_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A_ : List[Any] = self.token_map[placeholder_token] A_ : Optional[int] = tokens[: 1 + int(len(snake_case_ ) * prop_tokens_to_load )] if vector_shuffle: A_ : Optional[Any] = copy.copy(snake_case_ ) random.shuffle(snake_case_ ) A_ : List[str] = text.replace(snake_case_ , ' '.join(snake_case_ ) ) return text def __call__( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , ) def lowerCamelCase_ ( self , snake_case_ , *snake_case_ , snake_case_=False , snake_case_=1.0 , **snake_case_ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( snake_case_ , vector_shuffle=snake_case_ , prop_tokens_to_load=snake_case_ ) , *snake_case_ , **snake_case_ , )
286
0
def SCREAMING_SNAKE_CASE_ ( __A : list[int] , __A : str ) -> list[int]: """simple docstring""" a_ : Any = int(__A ) # Initialize Result a_ : Tuple = [] # Traverse through all denomination for denomination in reversed(__A ): # Find denominations while int(__A ) >= int(__A ): total_value -= int(__A ) answer.append(__A ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : Union[str, Any] = '0' if ( input('Do you want to enter your denominations ? (yY/n): ').strip().lower() == "y" ): UpperCAmelCase_ : List[Any] = int(input('Enter the number of denominations you want to add: ').strip()) for i in range(0, n): denominations.append(int(input(F'Denomination {i}: ').strip())) UpperCAmelCase_ : str = input('Enter the change you want to make in Indian Currency: ').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase_ : List[Any] = [1, 2, 5, 10, 20, 50, 100, 500, 2000] UpperCAmelCase_ : str = input('Enter the change you want to make: ').strip() if int(value) == 0 or int(value) < 0: print('The total value cannot be zero or negative.') else: print(F'Following is minimal change for {value}: ') UpperCAmelCase_ : Optional[Any] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=' ')
32
"""simple docstring""" def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) A_ : Any = hex_num[0] == '-' if is_negative: A_ : Optional[Any] = hex_num[1:] try: A_ : Tuple = int(_UpperCAmelCase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) A_ : Union[str, Any] = '' while int_num > 0: A_ : Optional[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
286
0
"""simple docstring""" import os # Precomputes a list of the 100 first triangular numbers __A : int = [int(0.5 * n * (n + 1)) for n in range(1, 101)] def lowercase ( ): lowercase_ : Optional[Any] = os.path.dirname(os.path.realpath(__snake_case ) ) lowercase_ : int = os.path.join(__snake_case , '''words.txt''' ) lowercase_ : Tuple = '''''' with open(__snake_case ) as f: lowercase_ : Union[str, Any] = f.readline() lowercase_ : Any = [word.strip('''"''' ) for word in words.strip('''\r\n''' ).split(''',''' )] lowercase_ : Optional[int] = [ word for word in [sum(ord(__snake_case ) - 6_4 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(__snake_case ) if __name__ == "__main__": print(solution())
33
"""simple docstring""" import qiskit def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Tuple = qiskit.Aer.get_backend('aer_simulator' ) A_ : str = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator A_ : Optional[Any] = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": lowerCamelCase_ : List[str] = half_adder(1, 1) print(F"Half Adder Output Qubit Counts: {counts}")
286
0
'''simple docstring''' import gc import unittest from diffusers import FlaxStableDiffusionInpaintPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class _a ( unittest.TestCase ): def A ( self : Union[str, Any] ): '''simple docstring''' super().tearDown() gc.collect() def A ( self : Tuple ): '''simple docstring''' UpperCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) UpperCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) UpperCAmelCase = '''xvjiarui/stable-diffusion-2-inpainting''' UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionInpaintPipeline.from_pretrained(lowercase , safety_checker=lowercase ) UpperCAmelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 50 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = num_samples * [init_image] UpperCAmelCase = num_samples * [mask_image] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = pipeline.prepare_inputs(lowercase , lowercase , lowercase ) # shard inputs and rng UpperCAmelCase = replicate(lowercase ) UpperCAmelCase = jax.random.split(lowercase , jax.device_count() ) UpperCAmelCase = shard(lowercase ) UpperCAmelCase = shard(lowercase ) UpperCAmelCase = shard(lowercase ) UpperCAmelCase = pipeline( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , jit=lowercase ) UpperCAmelCase = output.images.reshape(lowercase , 512 , 512 , 3 ) UpperCAmelCase = images[0, 253:256, 253:256, -1] UpperCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase = jnp.array( [0.361_1307, 0.3764_9736, 0.375_7408, 0.3821_3953, 0.3929_5167, 0.384_1631, 0.4155_4978, 0.413_7475, 0.421_7084] ) print(f"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
34
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Any = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Tuple = """xmod""" def __init__( self , snake_case_=3_0_5_2_2 , snake_case_=7_6_8 , snake_case_=1_2 , snake_case_=1_2 , snake_case_=3_0_7_2 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=2 , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=1 , snake_case_=0 , snake_case_=2 , snake_case_="absolute" , snake_case_=True , snake_case_=None , snake_case_=False , snake_case_=2 , snake_case_=False , snake_case_=True , snake_case_=True , snake_case_=("en_XX",) , snake_case_=None , **snake_case_ , ): """simple docstring""" super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) A_ : Union[str, Any] = vocab_size A_ : Any = hidden_size A_ : List[str] = num_hidden_layers A_ : Tuple = num_attention_heads A_ : int = hidden_act A_ : Any = intermediate_size A_ : Any = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Union[str, Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : List[str] = initializer_range A_ : Any = layer_norm_eps A_ : Optional[Any] = position_embedding_type A_ : int = use_cache A_ : Dict = classifier_dropout A_ : int = pre_norm A_ : Optional[Any] = adapter_reduction_factor A_ : List[Any] = adapter_layer_norm A_ : int = adapter_reuse_layer_norm A_ : Dict = ln_before_adapter A_ : List[str] = list(snake_case_ ) A_ : Union[str, Any] = default_language class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' @property def lowerCamelCase_ ( self ): """simple docstring""" if self.task == "multiple-choice": A_ : Dict = {0: 'batch', 1: 'choice', 2: 'sequence'} else: A_ : int = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
286
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __a = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["GLPNFeatureExtractor"] __a = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
35
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowercase_ : Dict = ["""image_processor""", """tokenizer"""] lowercase_ : Union[str, Any] = """ViltImageProcessor""" lowercase_ : Any = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): """simple docstring""" A_ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) A_ : Dict = kwargs.pop('feature_extractor' ) A_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) A_ : List[str] = self.image_processor def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): """simple docstring""" A_ : str = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel_values + pixel_mask A_ : Optional[int] = self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase_ ( self , *snake_case_ , **snake_case_ ): """simple docstring""" return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase_ ( self ): """simple docstring""" A_ : Any = self.tokenizer.model_input_names A_ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase_ ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
286
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { "asapp/sew-d-tiny-100k": "https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json", # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class UpperCAmelCase_ ( a): lowerCamelCase__ = 'sew-d' def __init__( self, __a=32, __a=768, __a=12, __a=12, __a=3072, __a=2, __a=512, __a=256, __a=True, __a=True, __a=("p2c", "c2p"), __a="layer_norm", __a="gelu_python", __a=0.1, __a=0.1, __a=0.1, __a=0.0, __a=0.1, __a=0.02, __a=1E-7, __a=1E-5, __a="group", __a="gelu", __a=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512), __a=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1), __a=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1), __a=False, __a=128, __a=16, __a=True, __a=0.05, __a=10, __a=2, __a=0.0, __a=10, __a=0, __a="mean", __a=False, __a=False, __a=256, __a=0, __a=1, __a=2, **__a, ): '''simple docstring''' super().__init__(**__a, pad_token_id=__a, bos_token_id=__a, eos_token_id=__a) _lowerCAmelCase : Tuple = hidden_size _lowerCAmelCase : List[Any] = feat_extract_norm _lowerCAmelCase : Dict = feat_extract_activation _lowerCAmelCase : Dict = list(__a) _lowerCAmelCase : Dict = list(__a) _lowerCAmelCase : List[str] = list(__a) _lowerCAmelCase : Dict = conv_bias _lowerCAmelCase : str = num_conv_pos_embeddings _lowerCAmelCase : int = num_conv_pos_embedding_groups _lowerCAmelCase : Dict = len(self.conv_dim) _lowerCAmelCase : List[str] = num_hidden_layers _lowerCAmelCase : List[Any] = intermediate_size _lowerCAmelCase : Dict = squeeze_factor _lowerCAmelCase : Any = max_position_embeddings _lowerCAmelCase : Optional[Any] = position_buckets _lowerCAmelCase : Optional[int] = share_att_key _lowerCAmelCase : str = relative_attention _lowerCAmelCase : Dict = norm_rel_ebd _lowerCAmelCase : Optional[int] = list(__a) _lowerCAmelCase : List[str] = hidden_act _lowerCAmelCase : Union[str, Any] = num_attention_heads _lowerCAmelCase : Optional[Any] = hidden_dropout _lowerCAmelCase : Dict = attention_dropout _lowerCAmelCase : str = activation_dropout _lowerCAmelCase : Optional[Any] = feat_proj_dropout _lowerCAmelCase : Tuple = final_dropout _lowerCAmelCase : Dict = layer_norm_eps _lowerCAmelCase : Union[str, Any] = feature_layer_norm_eps _lowerCAmelCase : str = initializer_range _lowerCAmelCase : str = vocab_size if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect." "It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`," f"but is `len(config.conv_dim) = {len(self.conv_dim)}`, `len(config.conv_stride)" f"= {len(self.conv_stride)}`, `len(config.conv_kernel) = {len(self.conv_kernel)}`.") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _lowerCAmelCase : List[Any] = apply_spec_augment _lowerCAmelCase : Optional[int] = mask_time_prob _lowerCAmelCase : Any = mask_time_length _lowerCAmelCase : Tuple = mask_time_min_masks _lowerCAmelCase : List[str] = mask_feature_prob _lowerCAmelCase : List[str] = mask_feature_length _lowerCAmelCase : Tuple = mask_feature_min_masks # ctc loss _lowerCAmelCase : Optional[Any] = ctc_loss_reduction _lowerCAmelCase : List[str] = ctc_zero_infinity # sequence classification _lowerCAmelCase : Optional[int] = use_weighted_layer_sum _lowerCAmelCase : Union[str, Any] = classifier_proj_size @property def snake_case__ ( self): '''simple docstring''' return functools.reduce(operator.mul, self.conv_stride, 1)
36
"""simple docstring""" from copy import deepcopy class _UpperCAmelCase : '''simple docstring''' def __init__( self , snake_case_ = None , snake_case_ = None ): """simple docstring""" if arr is None and size is not None: A_ : Union[str, Any] = size A_ : List[str] = [0] * size elif arr is not None: self.init(snake_case_ ) else: raise ValueError('Either arr or size must be specified' ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Union[str, Any] = len(snake_case_ ) A_ : Optional[int] = deepcopy(snake_case_ ) for i in range(1 , self.size ): A_ : Optional[Any] = self.next_(snake_case_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCamelCase_ ( self ): """simple docstring""" A_ : int = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): A_ : Optional[int] = self.next_(snake_case_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index + (index & (-index)) @staticmethod def lowerCamelCase_ ( snake_case_ ): """simple docstring""" return index - (index & (-index)) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value A_ : List[str] = self.next_(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" self.add(snake_case_ , value - self.get(snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" if right == 0: return 0 A_ : Any = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] A_ : Tuple = self.prev(snake_case_ ) return result def lowerCamelCase_ ( self , snake_case_ , snake_case_ ): """simple docstring""" return self.prefix(snake_case_ ) - self.prefix(snake_case_ ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" return self.query(snake_case_ , index + 1 ) def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" value -= self.tree[0] if value < 0: return -1 A_ : List[Any] = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 A_ : Tuple = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
286
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCAmelCase = { '''configuration_convnext''': ['''CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ConvNextConfig''', '''ConvNextOnnxConfig'''] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ['''ConvNextFeatureExtractor'''] _lowerCAmelCase = ['''ConvNextImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ '''CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ConvNextForImageClassification''', '''ConvNextModel''', '''ConvNextPreTrainedModel''', '''ConvNextBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ '''TFConvNextForImageClassification''', '''TFConvNextModel''', '''TFConvNextPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
37
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self , snake_case_ = 7_6_8 , ): """simple docstring""" super().__init__() A_ : Optional[int] = nn.Parameter(torch.zeros(1 , snake_case_ ) ) A_ : Optional[int] = nn.Parameter(torch.ones(1 , snake_case_ ) ) def lowerCamelCase_ ( self , snake_case_ = None , snake_case_ = None , ): """simple docstring""" A_ : str = nn.Parameter(self.mean.to(snake_case_ ).to(snake_case_ ) ) A_ : Optional[int] = nn.Parameter(self.std.to(snake_case_ ).to(snake_case_ ) ) return self def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : Tuple = (embeds - self.mean) * 1.0 / self.std return embeds def lowerCamelCase_ ( self , snake_case_ ): """simple docstring""" A_ : List[str] = (embeds * self.std) + self.mean return embeds
286
0
import warnings from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Union[str, Any] = ["""image_processor""", """tokenizer"""] snake_case__ : Optional[Any] = """FlavaImageProcessor""" snake_case__ : List[str] = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self : List[Any] , __lowerCamelCase : List[str]=None , __lowerCamelCase : Dict=None , **__lowerCamelCase : Tuple ): UpperCamelCase :Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __lowerCamelCase , ) UpperCamelCase :List[Any] = kwargs.pop("""feature_extractor""" ) UpperCamelCase :int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__lowerCamelCase , __lowerCamelCase ) UpperCamelCase :Optional[int] = self.image_processor def __call__( self : Optional[Any] , __lowerCamelCase : Optional[ImageInput] = None , __lowerCamelCase : Optional[Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]]] = None , __lowerCamelCase : bool = True , __lowerCamelCase : Union[bool, str, PaddingStrategy] = False , __lowerCamelCase : Union[bool, str, TruncationStrategy] = False , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = 0 , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = True , __lowerCamelCase : Optional[Union[str, TensorType]] = None , **__lowerCamelCase : int , ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCamelCase :List[str] = self.tokenizer( text=__lowerCamelCase , add_special_tokens=__lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=__lowerCamelCase , stride=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_token_type_ids=__lowerCamelCase , return_attention_mask=__lowerCamelCase , return_overflowing_tokens=__lowerCamelCase , return_special_tokens_mask=__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , return_length=__lowerCamelCase , verbose=__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase , ) if images is not None: UpperCamelCase :Dict = self.image_processor( __lowerCamelCase , return_image_mask=__lowerCamelCase , return_codebook_pixels=__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase , ) if text is not None and images is not None: encoding.update(__lowerCamelCase ) return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__lowerCamelCase ) , tensor_type=__lowerCamelCase ) def _A ( self : Any , *__lowerCamelCase : int , **__lowerCamelCase : Optional[int] ): return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase ) def _A ( self : Union[str, Any] , *__lowerCamelCase : Dict , **__lowerCamelCase : List[str] ): return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase ) @property def _A ( self : Tuple ): UpperCamelCase :Optional[Any] = self.tokenizer.model_input_names UpperCamelCase :Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _A ( self : Optional[Any] ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , __lowerCamelCase , ) return self.image_processor_class @property def _A ( self : List[Any] ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , __lowerCamelCase , ) return self.image_processor
38
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCamelCase_ : Any = HfArgumentParser(InitializationArguments) lowerCamelCase_ : Union[str, Any] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowerCamelCase_ : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowerCamelCase_ : Tuple = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) lowerCamelCase_ : int = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowerCamelCase_ : Any = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
286
0
import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal _a = datasets.utils.logging.get_logger(__name__) _a = ['''names''', '''prefix'''] _a = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] _a = ['''encoding_errors''', '''on_bad_lines'''] _a = ['''date_format'''] @dataclass class __lowerCamelCase ( datasets.BuilderConfig): """simple docstring""" UpperCamelCase__ = "," UpperCamelCase__ = None UpperCamelCase__ = "infer" UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = False UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = "." UpperCamelCase__ = None UpperCamelCase__ = '"' UpperCamelCase__ = 0 UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = 0 UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = None UpperCamelCase__ = 1_0000 UpperCamelCase__ = None UpperCamelCase__ = "strict" UpperCamelCase__ = "error" UpperCamelCase__ = None def UpperCamelCase ( self ): """simple docstring""" if self.delimiter is not None: _UpperCAmelCase = self.delimiter if self.column_names is not None: _UpperCAmelCase = self.column_names @property def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = { 'sep': self.sep, 'header': self.header, 'names': self.names, 'index_col': self.index_col, 'usecols': self.usecols, 'prefix': self.prefix, 'mangle_dupe_cols': self.mangle_dupe_cols, 'engine': self.engine, 'converters': self.converters, 'true_values': self.true_values, 'false_values': self.false_values, 'skipinitialspace': self.skipinitialspace, 'skiprows': self.skiprows, 'nrows': self.nrows, 'na_values': self.na_values, 'keep_default_na': self.keep_default_na, 'na_filter': self.na_filter, 'verbose': self.verbose, 'skip_blank_lines': self.skip_blank_lines, 'thousands': self.thousands, 'decimal': self.decimal, 'lineterminator': self.lineterminator, 'quotechar': self.quotechar, 'quoting': self.quoting, 'escapechar': self.escapechar, 'comment': self.comment, 'encoding': self.encoding, 'dialect': self.dialect, 'error_bad_lines': self.error_bad_lines, 'warn_bad_lines': self.warn_bad_lines, 'skipfooter': self.skipfooter, 'doublequote': self.doublequote, 'memory_map': self.memory_map, 'float_precision': self.float_precision, 'chunksize': self.chunksize, 'encoding_errors': self.encoding_errors, 'on_bad_lines': self.on_bad_lines, 'date_format': self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , UpperCAmelCase ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __lowerCamelCase ( datasets.ArrowBasedBuilder): """simple docstring""" UpperCamelCase__ = CsvConfig def UpperCamelCase ( self ): """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) _UpperCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase , (str, list, tuple) ): _UpperCAmelCase = data_files if isinstance(UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = [files] _UpperCAmelCase = [dl_manager.iter_files(UpperCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files} )] _UpperCAmelCase = [] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = [files] _UpperCAmelCase = [dl_manager.iter_files(UpperCAmelCase ) for file in files] splits.append(datasets.SplitGenerator(name=UpperCAmelCase , gen_kwargs={'files': files} ) ) return splits def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if self.config.features is not None: _UpperCAmelCase = self.config.features.arrow_schema if all(not require_storage_cast(UpperCAmelCase ) for feature in self.config.features.values() ): # cheaper cast _UpperCAmelCase = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=UpperCAmelCase ) else: # more expensive cast; allows str <-> int/float or str to Audio for example _UpperCAmelCase = table_cast(UpperCAmelCase , UpperCAmelCase ) return pa_table def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str _UpperCAmelCase = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(UpperCAmelCase ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase ) ): _UpperCAmelCase = pd.read_csv(UpperCAmelCase , iterator=UpperCAmelCase , dtype=UpperCAmelCase , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(UpperCAmelCase ): _UpperCAmelCase = pa.Table.from_pandas(UpperCAmelCase ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(UpperCAmelCase ) except ValueError as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase )}: {e}""" ) raise
39
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser lowerCamelCase_ : Any = re.compile(r'\s+') def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" return {"hash": hashlib.mda(re.sub(_UpperCAmelCase , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = [len(_UpperCAmelCase ) for line in example['content'].splitlines()] return {"line_mean": np.mean(_UpperCAmelCase ), "line_max": max(_UpperCAmelCase )} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 ): """simple docstring""" A_ : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] A_ : List[str] = example['content'].splitlines() for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=5 , _UpperCAmelCase=0.05 ): """simple docstring""" A_ : Any = ['unit tests', 'test file', 'configuration file'] A_ : Dict = example['content'].splitlines() A_ : List[Any] = 0 A_ : str = 0 # first test for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test A_ : Tuple = example['content'].count('\n' ) A_ : Tuple = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = ['def ', 'class ', 'for ', 'while '] A_ : Tuple = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=4 ): """simple docstring""" A_ : Union[str, Any] = example['content'].splitlines() A_ : Any = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = tokenizer(example['content'] , truncation=_UpperCAmelCase )['input_ids'] A_ : Dict = len(example['content'] ) / len(_UpperCAmelCase ) return {"ratio": ratio} def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : Any = {} results.update(get_hash(_UpperCAmelCase ) ) results.update(line_stats(_UpperCAmelCase ) ) results.update(alpha_stats(_UpperCAmelCase ) ) results.update(char_token_ratio(_UpperCAmelCase ) ) results.update(is_autogenerated(_UpperCAmelCase ) ) results.update(is_config_or_test(_UpperCAmelCase ) ) results.update(has_no_keywords(_UpperCAmelCase ) ) results.update(has_few_assignments(_UpperCAmelCase ) ) return results def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if not check_uniques(_UpperCAmelCase , _UpperCAmelCase ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" with open(_UpperCAmelCase , 'rb' ) as f_in: with gzip.open(str(_UpperCAmelCase ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) os.unlink(_UpperCAmelCase ) # Settings lowerCamelCase_ : Optional[int] = HfArgumentParser(PreprocessingArguments) lowerCamelCase_ : Optional[Any] = parser.parse_args() if args.num_workers is None: lowerCamelCase_ : int = multiprocessing.cpu_count() lowerCamelCase_ : Tuple = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset lowerCamelCase_ : Tuple = time.time() lowerCamelCase_ : Tuple = load_dataset(args.dataset_name, split='train') print(F"Time to load dataset: {time.time()-t_start:.2f}") # Run preprocessing lowerCamelCase_ : List[str] = time.time() lowerCamelCase_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(F"Time to preprocess dataset: {time.time()-t_start:.2f}") # Deduplicate hashes lowerCamelCase_ : int = set(ds.unique('hash')) lowerCamelCase_ : Union[str, Any] = len(uniques) / len(ds) print(F"Fraction of duplicates: {1-frac:.2%}") # Deduplicate data and apply heuristics lowerCamelCase_ : Optional[int] = time.time() lowerCamelCase_ : Tuple = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(F"Time to filter dataset: {time.time()-t_start:.2f}") print(F"Size of filtered dataset: {len(ds_filter)}") # Deduplicate with minhash and jaccard similarity if args.near_deduplication: lowerCamelCase_ : Union[str, Any] = time.time() lowerCamelCase_ , lowerCamelCase_ : str = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F"Time to deduplicate dataset: {time.time()-t_start:.2f}") print(F"Size of deduplicate dataset: {len(ds_filter)}") # Save data in batches of samples_per_file lowerCamelCase_ : Tuple = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) lowerCamelCase_ : Optional[Any] = output_dir / 'data' data_dir.mkdir(exist_ok=True) lowerCamelCase_ : List[str] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): lowerCamelCase_ : Optional[int] = str(data_dir / F"file-{file_number+1:012}.json") lowerCamelCase_ : List[str] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F"Time to save dataset: {time.time()-t_start:.2f}")
286
0
"""simple docstring""" import unittest from knapsack import knapsack as k class _A ( unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any]): a : str = 0 a : Optional[int] = [0] a : Union[str, Any] = [0] a : Any = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 0) a : List[str] = [60] a : str = [10] a : Optional[int] = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 0) def __snake_case ( self : Optional[int]): a : Any = 3 a : str = [1, 2, 3] a : Tuple = [3, 2, 1] a : Any = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 5) def __snake_case ( self : Tuple): a : int = 50 a : List[Any] = [60, 100, 120] a : Optional[int] = [10, 20, 30] a : str = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 220) if __name__ == "__main__": unittest.main()
40
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : Optional[Any] = [] # fmt: off # stem: rename_keys.append(('cls_token', 'vit.embeddings.cls_token') ) rename_keys.append(('pos_embed', 'vit.embeddings.position_embeddings') ) rename_keys.append(('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias') ) # backbone rename_keys.append(('patch_embed.backbone.stem.conv.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.weight', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight') ) rename_keys.append(('patch_embed.backbone.stem.norm.bias', 'vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((f"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", f"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A_ : List[str] = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) # fmt: on return rename_keys def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: A_ : List[str] = '' else: A_ : Dict = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ : str = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) A_ : List[Any] = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A_ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A_ : Tuple = in_proj_bias[: config.hidden_size] A_ : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A_ : Tuple = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ ( _UpperCAmelCase ): """simple docstring""" A_ : List[str] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = dct.pop(_UpperCAmelCase ) A_ : Optional[int] = val def UpperCAmelCase__ ( ): """simple docstring""" A_ : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' A_ : int = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ): """simple docstring""" A_ : List[Any] = BitConfig( global_padding='same' , layer_type='bottleneck' , depths=(3, 4, 9) , out_features=['stage3'] , embedding_dynamic_padding=_UpperCAmelCase , ) A_ : Optional[int] = ViTHybridConfig(backbone_config=_UpperCAmelCase , image_size=384 , num_labels=1000 ) A_ : Union[str, Any] = False # load original model from timm A_ : List[Any] = timm.create_model(_UpperCAmelCase , pretrained=_UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ : Tuple = timm_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) A_ : Any = create_rename_keys(_UpperCAmelCase , _UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) A_ : Union[str, Any] = 'huggingface/label-files' A_ : Dict = 'imagenet-1k-id2label.json' A_ : List[str] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) A_ : str = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} A_ : Any = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": A_ : List[Any] = ViTHybridModel(_UpperCAmelCase ).eval() else: A_ : str = ViTHybridForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # create image processor A_ : Dict = create_transform(**resolve_data_config({} , model=_UpperCAmelCase ) ) A_ : List[str] = transform.transforms A_ : List[str] = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } A_ : Tuple = ViTHybridImageProcessor( do_resize=_UpperCAmelCase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_UpperCAmelCase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=_UpperCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) A_ : Optional[Any] = prepare_img() A_ : Any = transform(_UpperCAmelCase ).unsqueeze(0 ) A_ : Dict = processor(_UpperCAmelCase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase ) # verify logits with torch.no_grad(): A_ : List[Any] = model(_UpperCAmelCase ) A_ : List[str] = outputs.logits print('Predicted class:' , logits.argmax(-1 ).item() ) if base_model: A_ : Union[str, Any] = timm_model.forward_features(_UpperCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_UpperCAmelCase , outputs.pooler_output , atol=1E-3 ) else: A_ : Tuple = timm_model(_UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase , outputs.logits , atol=1E-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(f"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: print(f"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(f"""ybelkada/{vit_name}""" ) processor.push_to_hub(f"""ybelkada/{vit_name}""" ) if __name__ == "__main__": lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_r50_s16_384', type=str, help='Name of the hybrid ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) lowerCamelCase_ : List[str] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
286
0