code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 1_6_0_0_0, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.feature_extraction_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) # load decoder from hub UpperCAmelCase = """hf-internal-testing/ngram-beam-search-decoder""" def _lowercase ( self : Optional[int] , **__lowerCamelCase : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.add_kwargs_tokens_map.copy() kwargs.update(__lowerCamelCase ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : List[str] , **__lowerCamelCase : str ) -> str: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : int ) -> Tuple: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **__lowerCamelCase ) def _lowercase ( self : Any ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __lowerCamelCase ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , __lowerCamelCase ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(__lowerCamelCase , """include""" ): WavaVecaProcessorWithLM( tokenizer=__lowerCamelCase , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = floats_list((3, 1_0_0_0) ) UpperCAmelCase = feature_extractor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(__lowerCamelCase , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = """This is a test string""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] , __lowerCamelCase : Optional[Any]=(2, 1_0, 1_6) , __lowerCamelCase : Union[str, Any]=7_7 ) -> List[str]: """simple docstring""" np.random.seed(__lowerCamelCase ) return np.random.rand(*__lowerCamelCase ) def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits(shape=(1_0, 1_6) , seed=1_3 ) UpperCAmelCase = processor.decode(__lowerCamelCase ) UpperCAmelCase = decoder.decode_beams(__lowerCamelCase )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def _lowercase ( self : List[Any] , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) else: with get_context(__lowerCamelCase ).Pool() as pool: UpperCAmelCase = processor.batch_decode(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = list(__lowerCamelCase ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase = decoder.decode_beams_batch(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(__lowerCamelCase , decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] , decoded_processor.text ) self.assertListEqual(__lowerCamelCase , decoded_processor.logit_score ) self.assertListEqual(__lowerCamelCase , decoded_processor.lm_score ) def _lowercase ( self : int ) -> int: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = 1_5 UpperCAmelCase = -20.0 UpperCAmelCase = -4.0 UpperCAmelCase = processor.batch_decode( __lowerCamelCase , beam_width=__lowerCamelCase , beam_prune_logp=__lowerCamelCase , token_min_logp=__lowerCamelCase , ) UpperCAmelCase = decoded_processor_out.text UpperCAmelCase = list(__lowerCamelCase ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase = decoder.decode_beams_batch( __lowerCamelCase , __lowerCamelCase , beam_width=__lowerCamelCase , beam_prune_logp=__lowerCamelCase , token_min_logp=__lowerCamelCase , ) UpperCAmelCase = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] , __lowerCamelCase ) self.assertTrue(np.array_equal(__lowerCamelCase , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , __lowerCamelCase , atol=1e-3 ) ) self.assertTrue(np.array_equal(__lowerCamelCase , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9_474] , __lowerCamelCase , atol=1e-3 ) ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = 2.0 UpperCAmelCase = 5.0 UpperCAmelCase = -20.0 UpperCAmelCase = True UpperCAmelCase = processor.batch_decode( __lowerCamelCase , alpha=__lowerCamelCase , beta=__lowerCamelCase , unk_score_offset=__lowerCamelCase , lm_score_boundary=__lowerCamelCase , ) UpperCAmelCase = decoded_processor_out.text UpperCAmelCase = list(__lowerCamelCase ) decoder.reset_params( alpha=__lowerCamelCase , beta=__lowerCamelCase , unk_score_offset=__lowerCamelCase , lm_score_boundary=__lowerCamelCase , ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase = decoder.decode_beams_batch( __lowerCamelCase , __lowerCamelCase , ) UpperCAmelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] , __lowerCamelCase ) UpperCAmelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained(__lowerCamelCase ) UpperCAmelCase = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = os.listdir(__lowerCamelCase ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = floats_list((3, 1_0_0_0) ) UpperCAmelCase = processor_wavaveca(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor_auto(__lowerCamelCase , return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2 ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = processor_wavaveca.batch_decode(__lowerCamelCase ) UpperCAmelCase = processor_auto.batch_decode(__lowerCamelCase ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , ) @staticmethod def _lowercase ( __lowerCamelCase : str , __lowerCamelCase : Tuple ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [d[key] for d in offsets] return retrieved_list def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = self._get_dummy_logits()[0] UpperCAmelCase = processor.decode(__lowerCamelCase , output_word_offsets=__lowerCamelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(__lowerCamelCase , __lowerCamelCase ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """end_offset""" ) , [1, 3, 5] ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = processor.batch_decode(__lowerCamelCase , output_word_offsets=__lowerCamelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(__lowerCamelCase , __lowerCamelCase ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(__lowerCamelCase , """word""" ) ) for o in outputs["""word_offsets"""]] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """end_offset""" ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" import torch UpperCAmelCase = load_dataset("""common_voice""" , """en""" , split="""train""" , streaming=__lowerCamelCase ) UpperCAmelCase = ds.cast_column("""audio""" , datasets.Audio(sampling_rate=1_6_0_0_0 ) ) UpperCAmelCase = iter(__lowerCamelCase ) UpperCAmelCase = next(__lowerCamelCase ) UpperCAmelCase = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase = processor(sample["""audio"""]["""array"""] , return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase = model(__lowerCamelCase ).logits.cpu().numpy() UpperCAmelCase = processor.decode(logits[0] , output_word_offsets=__lowerCamelCase ) UpperCAmelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(__lowerCamelCase , """word""" ) ) , __lowerCamelCase ) self.assertEqual(""" """.join(self.get_from_offsets(__lowerCamelCase , """word""" ) ) , output.text ) # output times UpperCAmelCase = torch.tensor(self.get_from_offsets(__lowerCamelCase , """start_time""" ) ) UpperCAmelCase = torch.tensor(self.get_from_offsets(__lowerCamelCase , """end_time""" ) ) # fmt: off UpperCAmelCase = torch.tensor([1.4_199, 1.6_599, 2.2_599, 3.0, 3.24, 3.5_999, 3.7_999, 4.0_999, 4.26, 4.94, 5.28, 5.6_599, 5.78, 5.94, 6.32, 6.5_399, 6.6_599] ) UpperCAmelCase = torch.tensor([1.5_399, 1.8_999, 2.9, 3.16, 3.5_399, 3.72, 4.0_199, 4.1_799, 4.76, 5.1_599, 5.5_599, 5.6_999, 5.86, 6.1_999, 6.38, 6.6_199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=0.01 ) ) self.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=0.01 ) )
701
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
0
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->int: if exponent == 1: return base if exponent % 2 == 0: UpperCAmelCase = _modexpt(lowerCAmelCase_ , exponent // 2 , lowerCAmelCase_ ) % modulo_value return (x * x) % modulo_value else: return (base * _modexpt(lowerCAmelCase_ , exponent - 1 , lowerCAmelCase_ )) % modulo_value def _UpperCamelCase ( lowerCAmelCase_ = 1_7_7_7 , lowerCAmelCase_ = 1_8_5_5 , lowerCAmelCase_ = 8 ) ->int: UpperCAmelCase = base for _ in range(1 , lowerCAmelCase_ ): UpperCAmelCase = _modexpt(lowerCAmelCase_ , lowerCAmelCase_ , 1_0**digits ) return result if __name__ == "__main__": print(F"""{solution() = }""")
702
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """roberta-base""": """https://huggingface.co/roberta-base/resolve/main/config.json""", """roberta-large""": """https://huggingface.co/roberta-large/resolve/main/config.json""", """roberta-large-mnli""": """https://huggingface.co/roberta-large-mnli/resolve/main/config.json""", """distilroberta-base""": """https://huggingface.co/distilroberta-base/resolve/main/config.json""", """roberta-base-openai-detector""": """https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json""", """roberta-large-openai-detector""": """https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''roberta''' def __init__( self : str , __lowerCamelCase : Optional[Any]=5_0_2_6_5 , __lowerCamelCase : Tuple=7_6_8 , __lowerCamelCase : Optional[Any]=1_2 , __lowerCamelCase : Tuple=1_2 , __lowerCamelCase : int=3_0_7_2 , __lowerCamelCase : Dict="gelu" , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : int=5_1_2 , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : Optional[Any]=1e-1_2 , __lowerCamelCase : List[Any]=1 , __lowerCamelCase : Tuple=0 , __lowerCamelCase : Tuple=2 , __lowerCamelCase : Optional[int]="absolute" , __lowerCamelCase : int=True , __lowerCamelCase : Tuple=None , **__lowerCamelCase : List[Any] , ) -> List[str]: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = position_embedding_type UpperCAmelCase = use_cache UpperCAmelCase = classifier_dropout class __lowercase ( __snake_case ): @property def _lowercase ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
703
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
0
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple=7 , __lowerCamelCase : Dict=3 , __lowerCamelCase : List[Any]=1_8 , __lowerCamelCase : Union[str, Any]=3_0 , __lowerCamelCase : Optional[int]=4_0_0 , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Any=True , ) -> str: """simple docstring""" UpperCAmelCase = size if size is not None else {"""height""": 1_8, """width""": 1_8} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = image_size UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_normalize def _lowercase ( self : Dict ) -> int: """simple docstring""" return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_866_443_634_033_203, 0.6_618_829_369_544_983, 0.3_891_746_401_786_804], [-0.6_042_559_146_881_104, -0.02_295_008_860_528_469, 0.5_423_797_369_003_296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = ImageGPTImageProcessor if is_vision_available() else None def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" UpperCAmelCase = ImageGPTImageProcessingTester(self ) @property def _lowercase ( self : Tuple ) -> Optional[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """clusters""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) def _lowercase ( self : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 1_8, """width""": 1_8} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 ) self.assertEqual(image_processor.size , {"""height""": 4_2, """width""": 4_2} ) def _lowercase ( self : Dict ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(__lowerCamelCase , obj[key] ) ) else: self.assertEqual(obj[key] , __lowerCamelCase ) def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase = os.path.join(__lowerCamelCase , """image_processor.json""" ) image_processor_first.to_json_file(__lowerCamelCase ) UpperCAmelCase = self.image_processing_class.from_json_file(__lowerCamelCase ).to_dict() UpperCAmelCase = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__lowerCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __lowerCamelCase ) def _lowercase ( self : str ) -> Any: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(__lowerCamelCase ) UpperCAmelCase = self.image_processing_class.from_pretrained(__lowerCamelCase ).to_dict() UpperCAmelCase = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__lowerCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __lowerCamelCase ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" pass def _UpperCamelCase ( ) ->Optional[Any]: UpperCAmelCase = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) UpperCAmelCase = Image.open(dataset[4]["""file"""] ) UpperCAmelCase = Image.open(dataset[5]["""file"""] ) UpperCAmelCase = [imagea, imagea] return images @require_vision @require_torch class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) UpperCAmelCase = prepare_images() # test non-batched UpperCAmelCase = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1_0_2_4) ) UpperCAmelCase = [3_0_6, 1_9_1, 1_9_1] self.assertEqual(encoding.input_ids[0, :3].tolist() , __lowerCamelCase ) # test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1_0_2_4) ) UpperCAmelCase = [3_0_3, 1_3, 1_3] self.assertEqual(encoding.input_ids[1, -3:].tolist() , __lowerCamelCase )
704
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
0
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __a = """\ @inproceedings{snover-etal-2006-study, title = \"A Study of Translation Edit Rate with Targeted Human Annotation\", author = \"Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John\", booktitle = \"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers\", month = aug # \" 8-12\", year = \"2006\", address = \"Cambridge, Massachusetts, USA\", publisher = \"Association for Machine Translation in the Americas\", url = \"https://aclanthology.org/2006.amta-papers.25\", pages = \"223--231\", } @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ __a = """\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. """ __a = """ Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: 'score' (float): TER score (num_edits / sum_ref_lengths * 100) 'num_edits' (int): The cumulative number of edits 'ref_length' (float): The cumulative average reference length Examples: Example 1: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 150.0, 'num_edits': 15, 'ref_length': 10.0} Example 2: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 62.5, 'num_edits': 5, 'ref_length': 8.0} Example 3: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {'score': 57.14285714285714, 'num_edits': 6, 'ref_length': 10.5} Example 4: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 0.0, 'num_edits': 0, 'ref_length': 8.0} Example 5: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 100.0, 'num_edits': 10, 'ref_length': 10.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""http://www.cs.umd.edu/~snover/tercom/""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#ter"""] , reference_urls=[ """https://github.com/jhclark/tercom""", ] , ) def _lowercase ( self : List[str] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , ) -> Tuple: """simple docstring""" UpperCAmelCase = len(references[0] ) if any(len(__lowerCamelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) UpperCAmelCase = [[refs[i] for refs in references] for i in range(__lowerCamelCase )] UpperCAmelCase = TER( normalized=__lowerCamelCase , no_punct=__lowerCamelCase , asian_support=__lowerCamelCase , case_sensitive=__lowerCamelCase , ) UpperCAmelCase = sb_ter.corpus_score(__lowerCamelCase , __lowerCamelCase ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
705
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowercase ( __snake_case ): UpperCamelCase = ['''image_processor''', '''tokenizer'''] UpperCamelCase = '''CLIPImageProcessor''' UpperCamelCase = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self : List[str] , __lowerCamelCase : Dict=None , __lowerCamelCase : Tuple=None , **__lowerCamelCase : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __lowerCamelCase , ) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__lowerCamelCase , __lowerCamelCase ) def __call__( self : Union[str, Any] , __lowerCamelCase : Dict=None , __lowerCamelCase : int=None , __lowerCamelCase : Tuple=None , **__lowerCamelCase : Optional[Any] ) -> Tuple: """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if images is not None: UpperCAmelCase = self.image_processor(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__lowerCamelCase ) , tensor_type=__lowerCamelCase ) def _lowercase ( self : List[str] , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Optional[Any] ) -> Tuple: """simple docstring""" return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : str , *__lowerCamelCase : Any , **__lowerCamelCase : Tuple ) -> int: """simple docstring""" return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase ) @property def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _lowercase ( self : str ) -> str: """simple docstring""" warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , __lowerCamelCase , ) return self.image_processor_class @property def _lowercase ( self : int ) -> Dict: """simple docstring""" warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , __lowerCamelCase , ) return self.image_processor
706
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
0
from collections.abc import Sequence def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = False ) ->float: if not arr: return 0 UpperCAmelCase = 0 if allow_empty_subarrays else float("""-inf""" ) UpperCAmelCase = 0.0 for num in arr: UpperCAmelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) UpperCAmelCase = max(lowerCAmelCase_ , lowerCAmelCase_ ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() __a = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F"""{max_subarray_sum(nums) = }""")
707
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
0
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = LongformerTokenizer UpperCamelCase = True UpperCamelCase = LongformerTokenizerFast UpperCamelCase = True def _lowercase ( self : Optional[Any] ) -> str: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) def _lowercase ( self : Dict , **__lowerCamelCase : Optional[Any] ) -> Dict: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Tuple , **__lowerCamelCase : Any ) -> Dict: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = """lower newer""" UpperCAmelCase = """lower newer""" return input_text, output_text def _lowercase ( self : Tuple ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCAmelCase = """lower newer""" UpperCAmelCase = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] UpperCAmelCase = tokenizer.tokenize(__lowerCamelCase ) # , add_prefix_space=True) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = tokens + [tokenizer.unk_token] UpperCAmelCase = [0, 1, 2, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() self.assertListEqual(tokenizer.encode("""Hello world!""" , add_special_tokens=__lowerCamelCase ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 2] ) self.assertListEqual( tokenizer.encode("""Hello world! cécé herlolip 418""" , add_special_tokens=__lowerCamelCase ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2] , ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.tokenizer_class.from_pretrained("""allenai/longformer-base-4096""" ) UpperCAmelCase = tokenizer.encode("""sequence builders""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode("""multi-sequence build""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode( """sequence builders""" , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode( """sequence builders""" , """multi-sequence build""" , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase , __lowerCamelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = """Encode this sequence.""" UpperCAmelCase = tokenizer.byte_encoder[""" """.encode("""utf-8""" )[0]] # Testing encoder arguments UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(__lowerCamelCase , __lowerCamelCase ) tokenizer.add_special_tokens({"""bos_token""": """<s>"""} ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(__lowerCamelCase , __lowerCamelCase ) # Testing spaces after special tokens UpperCAmelCase = """<mask>""" tokenizer.add_special_tokens( {"""mask_token""": AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase )} ) # mask token has a left space UpperCAmelCase = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) UpperCAmelCase = """Encode <mask> sequence""" UpperCAmelCase = """Encode <mask>sequence""" UpperCAmelCase = tokenizer.encode(__lowerCamelCase ) UpperCAmelCase = encoded.index(__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase ) UpperCAmelCase = encoded.index(__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" pass def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = self.tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = """A, <mask> AllenNLP sentence.""" UpperCAmelCase = tokenizer_r.encode_plus(__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_token_type_ids=__lowerCamelCase ) UpperCAmelCase = tokenizer_p.encode_plus(__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_token_type_ids=__lowerCamelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) UpperCAmelCase = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) UpperCAmelCase = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual( __lowerCamelCase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( __lowerCamelCase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) UpperCAmelCase = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["""add_prefix_space"""] , __lowerCamelCase ) self.assertEqual(post_processor_state["""add_prefix_space"""] , __lowerCamelCase ) self.assertEqual(post_processor_state["""trim_offsets"""] , __lowerCamelCase ) def _lowercase ( self : Any ) -> Tuple: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` UpperCAmelCase = F"""{text_of_1_token} {text_of_1_token}""" UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ) + 1, len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ) + 1, len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ), len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ), len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = F""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__lowerCamelCase ) + 1, 1 + len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__lowerCamelCase ), 1 + len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__lowerCamelCase ), 1 + len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , )
708
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __a = logging.get_logger(__name__) __a = { """SenseTime/deformable-detr""": """https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json""", # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class __lowercase ( __snake_case ): UpperCamelCase = '''deformable_detr''' UpperCamelCase = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self : int , __lowerCamelCase : Dict=True , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Dict=3 , __lowerCamelCase : Tuple=3_0_0 , __lowerCamelCase : str=1_0_2_4 , __lowerCamelCase : Optional[Any]=6 , __lowerCamelCase : List[Any]=1_0_2_4 , __lowerCamelCase : str=8 , __lowerCamelCase : Optional[Any]=6 , __lowerCamelCase : Tuple=1_0_2_4 , __lowerCamelCase : Tuple=8 , __lowerCamelCase : Tuple=0.0 , __lowerCamelCase : Any=True , __lowerCamelCase : int="relu" , __lowerCamelCase : List[str]=2_5_6 , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : Tuple=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : str=False , __lowerCamelCase : Tuple="sine" , __lowerCamelCase : Optional[Any]="resnet50" , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Dict=4 , __lowerCamelCase : str=4 , __lowerCamelCase : Union[str, Any]=4 , __lowerCamelCase : Dict=False , __lowerCamelCase : Union[str, Any]=3_0_0 , __lowerCamelCase : Tuple=False , __lowerCamelCase : Optional[Any]=1 , __lowerCamelCase : List[str]=5 , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : Tuple=1 , __lowerCamelCase : Optional[Any]=1 , __lowerCamelCase : Optional[int]=5 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : List[str]=0.25 , __lowerCamelCase : Any=False , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) UpperCAmelCase = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = backbone_config.get("""model_type""" ) UpperCAmelCase = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase = config_class.from_dict(__lowerCamelCase ) UpperCAmelCase = use_timm_backbone UpperCAmelCase = backbone_config UpperCAmelCase = num_channels UpperCAmelCase = num_queries UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = init_xavier_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = auxiliary_loss UpperCAmelCase = position_embedding_type UpperCAmelCase = backbone UpperCAmelCase = use_pretrained_backbone UpperCAmelCase = dilation # deformable attributes UpperCAmelCase = num_feature_levels UpperCAmelCase = encoder_n_points UpperCAmelCase = decoder_n_points UpperCAmelCase = two_stage UpperCAmelCase = two_stage_num_proposals UpperCAmelCase = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError("""If two_stage is True, with_box_refine must be True.""" ) # Hungarian matcher UpperCAmelCase = class_cost UpperCAmelCase = bbox_cost UpperCAmelCase = giou_cost # Loss coefficients UpperCAmelCase = mask_loss_coefficient UpperCAmelCase = dice_loss_coefficient UpperCAmelCase = bbox_loss_coefficient UpperCAmelCase = giou_loss_coefficient UpperCAmelCase = eos_coefficient UpperCAmelCase = focal_alpha UpperCAmelCase = disable_custom_kernels super().__init__(is_encoder_decoder=__lowerCamelCase , **__lowerCamelCase ) @property def _lowercase ( self : Any ) -> int: """simple docstring""" return self.encoder_attention_heads @property def _lowercase ( self : str ) -> int: """simple docstring""" return self.d_model def _lowercase ( self : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCAmelCase = self.backbone_config.to_dict() UpperCAmelCase = self.__class__.model_type return output
709
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
0
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __a = TypeVar("""T""") class __lowercase ( Generic[T] ): def __init__( self : Optional[int] , __lowerCamelCase : T ) -> str: """simple docstring""" UpperCAmelCase = data UpperCAmelCase = None def __str__( self : Dict ) -> str: """simple docstring""" return F"""{self.data}""" class __lowercase ( Generic[T] ): def __init__( self : List[Any] ) -> None: """simple docstring""" UpperCAmelCase = None def __iter__( self : Tuple ) -> Iterator[T]: """simple docstring""" UpperCAmelCase = self.top while node: yield node.data UpperCAmelCase = node.next def __str__( self : Union[str, Any] ) -> str: """simple docstring""" return "->".join([str(__lowerCamelCase ) for item in self] ) def __len__( self : Dict ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _lowercase ( self : Union[str, Any] ) -> bool: """simple docstring""" return self.top is None def _lowercase ( self : Any , __lowerCamelCase : T ) -> None: """simple docstring""" UpperCAmelCase = Node(__lowerCamelCase ) if not self.is_empty(): UpperCAmelCase = self.top UpperCAmelCase = node def _lowercase ( self : Dict ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , __lowerCamelCase ) UpperCAmelCase = self.top UpperCAmelCase = self.top.next return pop_node.data def _lowercase ( self : str ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _lowercase ( self : int ) -> None: """simple docstring""" UpperCAmelCase = None if __name__ == "__main__": from doctest import testmod testmod()
710
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
0
from ...configuration_utils import PretrainedConfig __a = { """google/tapas-base-finetuned-sqa""": ( """https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json""" ), """google/tapas-base-finetuned-wtq""": ( """https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json""" ), """google/tapas-base-finetuned-wikisql-supervised""": ( """https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json""" ), """google/tapas-base-finetuned-tabfact""": ( """https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json""" ), } class __lowercase ( __snake_case ): UpperCamelCase = '''tapas''' def __init__( self : str , __lowerCamelCase : Optional[Any]=3_0_5_2_2 , __lowerCamelCase : Dict=7_6_8 , __lowerCamelCase : List[Any]=1_2 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : str=3_0_7_2 , __lowerCamelCase : int="gelu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : int=0.1 , __lowerCamelCase : List[Any]=1_0_2_4 , __lowerCamelCase : Dict=[3, 2_5_6, 2_5_6, 2, 2_5_6, 2_5_6, 1_0] , __lowerCamelCase : Dict=0.02 , __lowerCamelCase : Union[str, Any]=1e-1_2 , __lowerCamelCase : Union[str, Any]=0 , __lowerCamelCase : int=10.0 , __lowerCamelCase : Optional[int]=0 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : Tuple=None , __lowerCamelCase : Any=1.0 , __lowerCamelCase : str=False , __lowerCamelCase : Dict=None , __lowerCamelCase : Any=1.0 , __lowerCamelCase : Tuple=1.0 , __lowerCamelCase : Any=False , __lowerCamelCase : List[str]=False , __lowerCamelCase : List[str]="ratio" , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=None , __lowerCamelCase : str=6_4 , __lowerCamelCase : List[Any]=3_2 , __lowerCamelCase : str=False , __lowerCamelCase : Dict=True , __lowerCamelCase : int=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : int=True , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Tuple=None , **__lowerCamelCase : Any , ) -> str: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_sizes UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps # Fine-tuning task hyperparameters UpperCAmelCase = positive_label_weight UpperCAmelCase = num_aggregation_labels UpperCAmelCase = aggregation_loss_weight UpperCAmelCase = use_answer_as_supervision UpperCAmelCase = answer_loss_importance UpperCAmelCase = use_normalized_answer_loss UpperCAmelCase = huber_loss_delta UpperCAmelCase = temperature UpperCAmelCase = aggregation_temperature UpperCAmelCase = use_gumbel_for_cells UpperCAmelCase = use_gumbel_for_aggregation UpperCAmelCase = average_approximation_function UpperCAmelCase = cell_selection_preference UpperCAmelCase = answer_loss_cutoff UpperCAmelCase = max_num_rows UpperCAmelCase = max_num_columns UpperCAmelCase = average_logits_per_cell UpperCAmelCase = select_one_column UpperCAmelCase = allow_empty_column_selection UpperCAmelCase = init_cell_selection_weights_to_zero UpperCAmelCase = reset_position_index_per_cell UpperCAmelCase = disable_per_token_loss # Aggregation hyperparameters UpperCAmelCase = aggregation_labels UpperCAmelCase = no_aggregation_label_index if isinstance(self.aggregation_labels , __lowerCamelCase ): UpperCAmelCase = {int(__lowerCamelCase ): v for k, v in aggregation_labels.items()}
711
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = IFInpaintingPipeline UpperCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} UpperCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS UpperCamelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} def _lowercase ( self : Any ) -> int: """simple docstring""" return self._get_dummy_components() def _lowercase ( self : Optional[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : str=0 ) -> Dict: """simple docstring""" if str(__lowerCamelCase ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(__lowerCamelCase ) else: UpperCAmelCase = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) UpperCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) UpperCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def _lowercase ( self : int ) -> List[str]: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1e-1 ) def _lowercase ( self : Dict ) -> List[str]: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" self._test_save_load_local() def _lowercase ( self : Tuple ) -> str: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
712
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
from __future__ import annotations import math def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->float: UpperCAmelCase = u for i in range(1 , lowerCAmelCase_ ): UpperCAmelCase = temp * (u - i) return temp def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""enter the numbers of values: """ ) ) UpperCAmelCase = [] for _ in range(lowerCAmelCase_ ): y.append([] ) for i in range(lowerCAmelCase_ ): for j in range(lowerCAmelCase_ ): y[i].append(lowerCAmelCase_ ) UpperCAmelCase = 0 print("""enter the values of parameters in a list: """ ) UpperCAmelCase = list(map(lowerCAmelCase_ , input().split() ) ) print("""enter the values of corresponding parameters: """ ) for i in range(lowerCAmelCase_ ): UpperCAmelCase = float(input() ) UpperCAmelCase = int(input("""enter the value to interpolate: """ ) ) UpperCAmelCase = (value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , lowerCAmelCase_ ): for j in range(n - i ): UpperCAmelCase = y[j + 1][i - 1] - y[j][i - 1] UpperCAmelCase = y[0][0] for i in range(1 , lowerCAmelCase_ ): summ += (ucal(lowerCAmelCase_ , lowerCAmelCase_ ) * y[0][i]) / math.factorial(lowerCAmelCase_ ) print(F"""the value at {value} is {summ}""" ) if __name__ == "__main__": main()
713
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
0
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) UpperCAmelCase = str(bin(lowerCAmelCase_ ) )[2:] # remove the leading "0b" UpperCAmelCase = str(bin(lowerCAmelCase_ ) )[2:] # remove the leading "0b" UpperCAmelCase = max(len(lowerCAmelCase_ ) , len(lowerCAmelCase_ ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(lowerCAmelCase_ ) , b_binary.zfill(lowerCAmelCase_ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
714
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = {"""vocab_file""": """spm_char.model"""} __a = { """vocab_file""": { """microsoft/speecht5_asr""": """https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model""", """microsoft/speecht5_tts""": """https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model""", """microsoft/speecht5_vc""": """https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model""", } } __a = { """microsoft/speecht5_asr""": 1024, """microsoft/speecht5_tts""": 1024, """microsoft/speecht5_vc""": 1024, } class __lowercase ( __snake_case ): UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : List[str]="<s>" , __lowerCamelCase : Tuple="</s>" , __lowerCamelCase : Any="<unk>" , __lowerCamelCase : Union[str, Any]="<pad>" , __lowerCamelCase : Optional[Dict[str, Any]] = None , **__lowerCamelCase : Any , ) -> None: """simple docstring""" UpperCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , unk_token=__lowerCamelCase , pad_token=__lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCamelCase , ) UpperCAmelCase = vocab_file UpperCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__lowerCamelCase ) @property def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" return self.sp_model.get_piece_size() def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = {self.convert_ids_to_tokens(__lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.__dict__.copy() UpperCAmelCase = None return state def __setstate__( self : int , __lowerCamelCase : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCAmelCase = {} UpperCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : str , __lowerCamelCase : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__lowerCamelCase , out_type=__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict ) -> Tuple: """simple docstring""" return self.sp_model.piece_to_id(__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.sp_model.IdToPiece(__lowerCamelCase ) return token def _lowercase ( self : Optional[int] , __lowerCamelCase : int ) -> List[str]: """simple docstring""" UpperCAmelCase = [] UpperCAmelCase = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__lowerCamelCase ) + token UpperCAmelCase = [] else: current_sub_tokens.append(__lowerCamelCase ) out_string += self.sp_model.decode(__lowerCamelCase ) return out_string.strip() def _lowercase ( self : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[Any]=None ) -> List[int]: """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _lowercase ( self : Dict , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None , __lowerCamelCase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) UpperCAmelCase = [1] if token_ids_a is None: return ([0] * len(__lowerCamelCase )) + suffix_ones return ([0] * len(__lowerCamelCase )) + ([0] * len(__lowerCamelCase )) + suffix_ones def _lowercase ( self : Tuple , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__lowerCamelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase = os.path.join( __lowerCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__lowerCamelCase , """wb""" ) as fi: UpperCAmelCase = self.sp_model.serialized_model_proto() fi.write(__lowerCamelCase ) return (out_vocab_file,)
715
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
0
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A ( __snake_case ): def __init__( self : Optional[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Dict ) -> Dict: """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM UpperCAmelCase = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=__lowerCamelCase , scheduler=__lowerCamelCase ) @torch.no_grad() def __call__( self : str , __lowerCamelCase : int = 1 , __lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __lowerCamelCase : float = 0.0 , __lowerCamelCase : int = 5_0 , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[str] = "pil" , __lowerCamelCase : bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" if isinstance(self.unet.config.sample_size , __lowerCamelCase ): UpperCAmelCase = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: UpperCAmelCase = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(__lowerCamelCase ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(__lowerCamelCase )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) UpperCAmelCase = randn_tensor(__lowerCamelCase , generator=__lowerCamelCase , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__lowerCamelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output UpperCAmelCase = self.unet(__lowerCamelCase , __lowerCamelCase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 UpperCAmelCase = self.scheduler.step( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , eta=__lowerCamelCase , use_clipped_model_output=__lowerCamelCase , generator=__lowerCamelCase ).prev_sample UpperCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase = self.numpy_to_pil(__lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__lowerCamelCase )
716
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
0
import math def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0 ) ->int: UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
717
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
0
import math def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return math.sqrt(lowerCAmelCase_ ) * math.sqrt(lowerCAmelCase_ ) == num def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
718
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy __a = logging.get_logger(__name__) class __lowercase ( __snake_case ): def __init__( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float , **__lowerCamelCase : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase = feature_size UpperCAmelCase = sampling_rate UpperCAmelCase = padding_value UpperCAmelCase = kwargs.pop("""padding_side""" , """right""" ) UpperCAmelCase = kwargs.pop("""return_attention_mask""" , __lowerCamelCase ) super().__init__(**__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] , __lowerCamelCase : Union[bool, str, PaddingStrategy] = True , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , ) -> BatchFeature: """simple docstring""" if isinstance(__lowerCamelCase , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( """You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`""" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase = processed_features[self.model_input_names[0]] UpperCAmelCase = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(__lowerCamelCase ) == 0: if return_attention_mask: UpperCAmelCase = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase = required_input[0] if isinstance(__lowerCamelCase , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase = 0 while len(required_input[index] ) == 0: index += 1 if index < len(__lowerCamelCase ): UpperCAmelCase = required_input[index][0] if return_tensors is None: if is_tf_tensor(__lowerCamelCase ): UpperCAmelCase = """tf""" elif is_torch_tensor(__lowerCamelCase ): UpperCAmelCase = """pt""" elif isinstance(__lowerCamelCase , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase = """np""" else: raise ValueError( F"""type of {first_element} unknown: {type(__lowerCamelCase )}. """ """Should be one of a python, numpy, pytorch or tensorflow object.""" ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase = to_numpy(__lowerCamelCase ) else: UpperCAmelCase = [to_numpy(__lowerCamelCase ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase = self._get_padding_strategies(padding=__lowerCamelCase , max_length=__lowerCamelCase ) UpperCAmelCase = processed_features[self.model_input_names[0]] UpperCAmelCase = len(__lowerCamelCase ) if not all(len(__lowerCamelCase ) == batch_size for v in processed_features.values() ): raise ValueError("""Some items in the output dictionary have a different batch size than others.""" ) UpperCAmelCase = [] for i in range(__lowerCamelCase ): UpperCAmelCase = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase = self._truncate( __lowerCamelCase , max_length=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , truncation=__lowerCamelCase , ) truncated_inputs.append(__lowerCamelCase ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase = PaddingStrategy.MAX_LENGTH UpperCAmelCase = {} for i in range(__lowerCamelCase ): # padding UpperCAmelCase = self._pad( truncated_inputs[i] , max_length=__lowerCamelCase , padding_strategy=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_attention_mask=__lowerCamelCase , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase = value.astype(np.floataa ) batch_outputs[key].append(__lowerCamelCase ) return BatchFeature(__lowerCamelCase , tensor_type=__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Union[Dict[str, np.ndarray], BatchFeature] , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , ) -> dict: """simple docstring""" UpperCAmelCase = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase = len(__lowerCamelCase ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(__lowerCamelCase ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase = np.ones(len(__lowerCamelCase ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase = max_length - len(__lowerCamelCase ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase = np.pad( processed_features["""attention_mask"""] , (0, difference) ) UpperCAmelCase = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase = np.pad( __lowerCamelCase , __lowerCamelCase , """constant""" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase = np.pad( processed_features["""attention_mask"""] , (difference, 0) ) UpperCAmelCase = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase = np.pad( __lowerCamelCase , __lowerCamelCase , """constant""" , constant_values=self.padding_value ) else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return processed_features def _lowercase ( self : int , __lowerCamelCase : Union[Dict[str, np.ndarray], BatchFeature] , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , ) -> Dict: """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("""When setting ``truncation=True``, make sure that ``max_length`` is defined.""" ) UpperCAmelCase = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase = len(__lowerCamelCase ) > max_length if needs_to_be_truncated: UpperCAmelCase = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase = processed_features["""attention_mask"""][:max_length] return processed_features def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Any=False , __lowerCamelCase : Any=None ) -> Dict: """simple docstring""" if padding is not False: if padding is True: UpperCAmelCase = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = PaddingStrategy(__lowerCamelCase ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = padding else: UpperCAmelCase = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( """Asking to pad but the feature_extractor does not have a padding value. Please select a value to use""" """ as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.""" ) return padding_strategy
719
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple=7 , __lowerCamelCase : int=3 , __lowerCamelCase : List[str]=1_0 , __lowerCamelCase : Any=1_8 , __lowerCamelCase : str=3_0 , __lowerCamelCase : Dict=4_0_0 , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Any=None , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=[0.5, 0.5, 0.5] , __lowerCamelCase : Any=[0.5, 0.5, 0.5] , __lowerCamelCase : str=None , ) -> List[Any]: """simple docstring""" UpperCAmelCase = size if size is not None else {"""shortest_edge""": 1_8} UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 1_8, """width""": 1_8} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = num_frames UpperCAmelCase = image_size UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = crop_size def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = VivitImageProcessor if is_vision_available() else None def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = VivitImageProcessingTester(self ) @property def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 1_8} ) self.assertEqual(image_processor.crop_size , {"""height""": 1_8, """width""": 1_8} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 4_2} ) self.assertEqual(image_processor.crop_size , {"""height""": 8_4, """width""": 8_4} ) def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCAmelCase = prepare_video_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase ) for video in video_inputs: self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _lowercase ( self : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_video_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , numpify=__lowerCamelCase ) for video in video_inputs: self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _lowercase ( self : List[Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_video_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , torchify=__lowerCamelCase ) for video in video_inputs: self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
720
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
0
'''simple docstring''' from string import ascii_uppercase __a = {str(ord(c) - 55): c for c in ascii_uppercase} def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""int() can't convert non-string with explicit base""" ) if num < 0: raise ValueError("""parameter must be positive int""" ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""'str' object cannot be interpreted as an integer""" ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""'float' object cannot be interpreted as an integer""" ) if base in (0, 1): raise ValueError("""base must be >= 2""" ) if base > 3_6: raise ValueError("""base must be <= 36""" ) UpperCAmelCase = """""" UpperCAmelCase = 0 UpperCAmelCase = 0 while div != 1: UpperCAmelCase , UpperCAmelCase = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) if base >= 1_1 and 9 < mod < 3_6: UpperCAmelCase = ALPHABET_VALUES[str(lowerCAmelCase_ )] else: UpperCAmelCase = str(lowerCAmelCase_ ) new_value += actual_value UpperCAmelCase = num // base UpperCAmelCase = div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(lowerCAmelCase_ ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
721
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
0
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def _UpperCamelCase ( lowerCAmelCase_ ) ->Dict: # A local function to see if a dot lands in the circle. def is_in_circle(lowerCAmelCase_ , lowerCAmelCase_ ) -> bool: UpperCAmelCase = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle UpperCAmelCase = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(lowerCAmelCase_ ) ) # The ratio of the area for circle to square is pi/4. UpperCAmelCase = proportion * 4 print(F"""The estimated value of pi is {pi_estimate}""" ) print(F"""The numpy value of pi is {pi}""" ) print(F"""The total error is {abs(pi - pi_estimate )}""" ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = 1.0 , ) ->float: return mean( function_to_integrate(uniform(lowerCAmelCase_ , lowerCAmelCase_ ) ) for _ in range(lowerCAmelCase_ ) ) * (max_value - min_value) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = 1.0 ) ->None: def identity_function(lowerCAmelCase_ ) -> float: return x UpperCAmelCase = area_under_curve_estimator( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = (max_value * max_value - min_value * min_value) / 2 print("""******************""" ) print(F"""Estimating area under y=x where x varies from {min_value} to {max_value}""" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {expected_value}""" ) print(F"""Total error is {abs(estimated_value - expected_value )}""" ) print("""******************""" ) def _UpperCamelCase ( lowerCAmelCase_ ) ->None: def function_to_integrate(lowerCAmelCase_ ) -> float: return sqrt(4.0 - x * x ) UpperCAmelCase = area_under_curve_estimator( lowerCAmelCase_ , lowerCAmelCase_ , 0.0 , 2.0 ) print("""******************""" ) print("""Estimating pi using area_under_curve_estimator""" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {pi}""" ) print(F"""Total error is {abs(estimated_value - pi )}""" ) print("""******************""" ) if __name__ == "__main__": import doctest doctest.testmod()
700
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings __a = R""" [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `\" / \"`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `\" // \"`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `\"wiki_dpr\"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `\"train\"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `\"compressed\"`) The index name of the index associated with the `dataset`. One can choose between `\"legacy\"`, `\"exact\"` and `\"compressed\"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a \"dummy\" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. """ @add_start_docstrings(__snake_case ) class __lowercase ( __snake_case ): UpperCamelCase = '''rag''' UpperCamelCase = True def __init__( self : List[str] , __lowerCamelCase : Dict=None , __lowerCamelCase : Tuple=True , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Dict=None , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : str=None , __lowerCamelCase : Dict=None , __lowerCamelCase : Tuple=" / " , __lowerCamelCase : List[str]=" // " , __lowerCamelCase : Union[str, Any]=5 , __lowerCamelCase : Tuple=3_0_0 , __lowerCamelCase : Any=7_6_8 , __lowerCamelCase : Dict=8 , __lowerCamelCase : Optional[int]="wiki_dpr" , __lowerCamelCase : int="train" , __lowerCamelCase : Optional[Any]="compressed" , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : List[Any]=None , __lowerCamelCase : str=False , __lowerCamelCase : Dict=False , __lowerCamelCase : int=0.0 , __lowerCamelCase : Dict=True , __lowerCamelCase : Dict=False , __lowerCamelCase : Union[str, Any]=False , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : int=None , **__lowerCamelCase : Optional[int] , ) -> Optional[Any]: """simple docstring""" super().__init__( bos_token_id=__lowerCamelCase , pad_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , forced_eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , prefix=__lowerCamelCase , vocab_size=__lowerCamelCase , **__lowerCamelCase , ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" UpperCAmelCase = kwargs.pop("""question_encoder""" ) UpperCAmelCase = question_encoder_config.pop("""model_type""" ) UpperCAmelCase = kwargs.pop("""generator""" ) UpperCAmelCase = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig UpperCAmelCase = AutoConfig.for_model(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = AutoConfig.for_model(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = reduce_loss UpperCAmelCase = label_smoothing UpperCAmelCase = exclude_bos_score UpperCAmelCase = do_marginalize UpperCAmelCase = title_sep UpperCAmelCase = doc_sep UpperCAmelCase = n_docs UpperCAmelCase = max_combined_length UpperCAmelCase = dataset UpperCAmelCase = dataset_split UpperCAmelCase = index_name UpperCAmelCase = retrieval_vector_size UpperCAmelCase = retrieval_batch_size UpperCAmelCase = passages_path UpperCAmelCase = index_path UpperCAmelCase = use_dummy_dataset UpperCAmelCase = output_retrieved UpperCAmelCase = do_deduplication UpperCAmelCase = use_cache if self.forced_eos_token_id is None: UpperCAmelCase = getattr(self.generator , """forced_eos_token_id""" , __lowerCamelCase ) @classmethod def _lowercase ( cls : Any , __lowerCamelCase : PretrainedConfig , __lowerCamelCase : PretrainedConfig , **__lowerCamelCase : Union[str, Any] ) -> PretrainedConfig: """simple docstring""" return cls(question_encoder=question_encoder_config.to_dict() , generator=generator_config.to_dict() , **__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = copy.deepcopy(self.__dict__ ) UpperCAmelCase = self.question_encoder.to_dict() UpperCAmelCase = self.generator.to_dict() UpperCAmelCase = self.__class__.model_type return output
701
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
0
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
702
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
0
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging __a = """\ """ __a = """ Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity """ __a = """ Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to 'cuda' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"] >>> results = perplexity.compute(model_id='gpt2', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 78.22 >>> print(round(results[\"perplexities\"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = datasets.load_dataset(\"wikitext\", ... \"wikitext-2-raw-v1\", ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=''] >>> results = perplexity.compute(model_id='gpt2', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 60.35 >>> print(round(results[\"perplexities\"][0], 2)) 81.12 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : List[str] ) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """input_texts""": datasets.Value("""string""" ), } ) , reference_urls=["""https://huggingface.co/docs/transformers/perplexity"""] , ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Tuple , __lowerCamelCase : int = 1_6 , __lowerCamelCase : bool = True , __lowerCamelCase : str=None ) -> Tuple: """simple docstring""" if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCAmelCase = """cuda""" else: UpperCAmelCase = """cuda""" if torch.cuda.is_available() else """cpu""" UpperCAmelCase = AutoModelForCausalLM.from_pretrained(__lowerCamelCase ) UpperCAmelCase = model.to(__lowerCamelCase ) UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCAmelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(__lowerCamelCase ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"""pad_token""": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCAmelCase = model.config.max_length - 1 else: UpperCAmelCase = model.config.max_length UpperCAmelCase = tokenizer( __lowerCamelCase , add_special_tokens=__lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=__lowerCamelCase , return_tensors="""pt""" , return_attention_mask=__lowerCamelCase , ).to(__lowerCamelCase ) UpperCAmelCase = encodings["""input_ids"""] UpperCAmelCase = encodings["""attention_mask"""] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCAmelCase = [] UpperCAmelCase = CrossEntropyLoss(reduction="""none""" ) for start_index in logging.tqdm(range(0 , len(__lowerCamelCase ) , __lowerCamelCase ) ): UpperCAmelCase = min(start_index + batch_size , len(__lowerCamelCase ) ) UpperCAmelCase = encoded_texts[start_index:end_index] UpperCAmelCase = attn_masks[start_index:end_index] if add_start_token: UpperCAmelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(__lowerCamelCase ) UpperCAmelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCAmelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(__lowerCamelCase ), attn_mask] , dim=1 ) UpperCAmelCase = encoded_batch with torch.no_grad(): UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase ).logits UpperCAmelCase = out_logits[..., :-1, :].contiguous() UpperCAmelCase = labels[..., 1:].contiguous() UpperCAmelCase = attn_mask[..., 1:].contiguous() UpperCAmelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , __lowerCamelCase ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(__lowerCamelCase )}
703
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
0
from random import shuffle import tensorflow as tf from numpy import array def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: UpperCAmelCase = int(lowerCAmelCase_ ) assert noofclusters < len(lowerCAmelCase_ ) # Find out the dimensionality UpperCAmelCase = len(vectors[0] ) # Will help select random centroids from among the available vectors UpperCAmelCase = list(range(len(lowerCAmelCase_ ) ) ) shuffle(lowerCAmelCase_ ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. UpperCAmelCase = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION UpperCAmelCase = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points UpperCAmelCase = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(lowerCAmelCase_ ) ] ##These nodes will assign the centroid Variables the appropriate ##values UpperCAmelCase = tf.placeholder("""float64""" , [dim] ) UpperCAmelCase = [] for centroid in centroids: cent_assigns.append(tf.assign(lowerCAmelCase_ , lowerCAmelCase_ ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) UpperCAmelCase = [tf.Variable(0 ) for i in range(len(lowerCAmelCase_ ) )] ##These nodes will assign an assignment Variable the appropriate ##value UpperCAmelCase = tf.placeholder("""int32""" ) UpperCAmelCase = [] for assignment in assignments: cluster_assigns.append(tf.assign(lowerCAmelCase_ , lowerCAmelCase_ ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input UpperCAmelCase = tf.placeholder("""float""" , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors UpperCAmelCase = tf.reduce_mean(lowerCAmelCase_ , 0 ) ##Node for computing Euclidean distances # Placeholders for input UpperCAmelCase = tf.placeholder("""float""" , [dim] ) UpperCAmelCase = tf.placeholder("""float""" , [dim] ) UpperCAmelCase = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(lowerCAmelCase_ , lowerCAmelCase_ ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input UpperCAmelCase = tf.placeholder("""float""" , [noofclusters] ) UpperCAmelCase = tf.argmin(lowerCAmelCase_ , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. UpperCAmelCase = tf.initialize_all_variables() # Initialize all variables sess.run(lowerCAmelCase_ ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. UpperCAmelCase = 1_0_0 for _ in range(lowerCAmelCase_ ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. UpperCAmelCase = [ sess.run(lowerCAmelCase_ , feed_dict={va: vect, va: sess.run(lowerCAmelCase_ )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input UpperCAmelCase = sess.run( lowerCAmelCase_ , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(lowerCAmelCase_ ): # Collect all the vectors assigned to this cluster UpperCAmelCase = [ vectors[i] for i in range(len(lowerCAmelCase_ ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location UpperCAmelCase = sess.run( lowerCAmelCase_ , feed_dict={mean_input: array(lowerCAmelCase_ )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments UpperCAmelCase = sess.run(lowerCAmelCase_ ) UpperCAmelCase = sess.run(lowerCAmelCase_ ) return centroids, assignments
704
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
0
'''simple docstring''' from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING __a = logging.get_logger(__name__) @add_end_docstrings(__snake_case ) class __lowercase ( __snake_case ): def __init__( self : Union[str, Any] , **__lowerCamelCase : Dict ) -> Optional[int]: """simple docstring""" super().__init__(**__lowerCamelCase ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F"""The {self.__class__} is only available in PyTorch.""" ) self.check_model_type(__lowerCamelCase ) def _lowercase ( self : Tuple , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = {} UpperCAmelCase = {} UpperCAmelCase = {} # preprocess args if "points_per_batch" in kwargs: UpperCAmelCase = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: UpperCAmelCase = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: UpperCAmelCase = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: UpperCAmelCase = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: UpperCAmelCase = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: UpperCAmelCase = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: UpperCAmelCase = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: UpperCAmelCase = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: UpperCAmelCase = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: UpperCAmelCase = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: UpperCAmelCase = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: UpperCAmelCase = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self : Any , __lowerCamelCase : Any , *__lowerCamelCase : Optional[int] , __lowerCamelCase : Any=None , __lowerCamelCase : Union[str, Any]=None , **__lowerCamelCase : List[str] ) -> Optional[Any]: """simple docstring""" return super().__call__(__lowerCamelCase , *__lowerCamelCase , num_workers=__lowerCamelCase , batch_size=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple=6_4 , __lowerCamelCase : int = 0 , __lowerCamelCase : float = 5_1_2 / 1_5_0_0 , __lowerCamelCase : Optional[int] = 3_2 , __lowerCamelCase : Optional[int] = 1 , ) -> Optional[int]: """simple docstring""" UpperCAmelCase = load_image(__lowerCamelCase ) UpperCAmelCase = self.image_processor.size["""longest_edge"""] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.image_processor.generate_crop_boxes( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self.image_processor(images=__lowerCamelCase , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": UpperCAmelCase = self.get_inference_context() with inference_context(): UpperCAmelCase = self._ensure_tensor_on_device(__lowerCamelCase , device=self.device ) UpperCAmelCase = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) UpperCAmelCase = image_embeddings UpperCAmelCase = grid_points.shape[1] UpperCAmelCase = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , __lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = grid_points[:, i : i + points_per_batch, :, :] UpperCAmelCase = input_labels[:, i : i + points_per_batch] UpperCAmelCase = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def _lowercase ( self : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Optional[int]=0.88 , __lowerCamelCase : List[Any]=0.95 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : str=1 , ) -> str: """simple docstring""" UpperCAmelCase = model_inputs.pop("""input_boxes""" ) UpperCAmelCase = model_inputs.pop("""is_last""" ) UpperCAmelCase = model_inputs.pop("""original_sizes""" ).tolist() UpperCAmelCase = model_inputs.pop("""reshaped_input_sizes""" ).tolist() UpperCAmelCase = self.model(**__lowerCamelCase ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks UpperCAmelCase = model_outputs["""pred_masks"""] UpperCAmelCase = self.image_processor.post_process_masks( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , binarize=__lowerCamelCase ) UpperCAmelCase = model_outputs["""iou_scores"""] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def _lowercase ( self : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : List[str]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : Optional[int]=0.7 , ) -> Tuple: """simple docstring""" UpperCAmelCase = [] UpperCAmelCase = [] UpperCAmelCase = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) UpperCAmelCase = torch.cat(__lowerCamelCase ) UpperCAmelCase = torch.cat(__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.image_processor.post_process_for_mask_generation( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = defaultdict(__lowerCamelCase ) for output in model_outputs: for k, v in output.items(): extra[k].append(__lowerCamelCase ) UpperCAmelCase = {} if output_rle_mask: UpperCAmelCase = rle_mask if output_bboxes_mask: UpperCAmelCase = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
705
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
0
'''simple docstring''' import os __a = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1000} def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 while index < len(lowerCAmelCase_ ) - 1: UpperCAmelCase = SYMBOLS[numerals[index]] UpperCAmelCase = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = """""" UpperCAmelCase = num // 1_0_0_0 numerals += m_count * "M" num %= 1_0_0_0 UpperCAmelCase = num // 1_0_0 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_0_0 UpperCAmelCase = num // 1_0 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 1_0 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def _UpperCamelCase ( lowerCAmelCase_ = "/p089_roman.txt" ) ->int: UpperCAmelCase = 0 with open(os.path.dirname(lowerCAmelCase_ ) + roman_numerals_filename ) as filea: UpperCAmelCase = filea.readlines() for line in lines: UpperCAmelCase = line.strip() UpperCAmelCase = parse_roman_numerals(lowerCAmelCase_ ) UpperCAmelCase = generate_roman_numerals(lowerCAmelCase_ ) savings += len(lowerCAmelCase_ ) - len(lowerCAmelCase_ ) return savings if __name__ == "__main__": print(F"""{solution() = }""")
706
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
0
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __a = logging.get_logger(__name__) class __lowercase ( __snake_case ): UpperCamelCase = '''AutoTokenizer''' UpperCamelCase = ['''tokenizer'''] UpperCamelCase = { '''semantic_prompt''': 1, '''coarse_prompt''': 2, '''fine_prompt''': 2, } def __init__( self : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any=None ) -> Any: """simple docstring""" super().__init__(__lowerCamelCase ) UpperCAmelCase = speaker_embeddings @classmethod def _lowercase ( cls : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : str="speaker_embeddings_path.json" , **__lowerCamelCase : Any ) -> str: """simple docstring""" if speaker_embeddings_dict_path is not None: UpperCAmelCase = get_file_from_repo( __lowerCamelCase , __lowerCamelCase , subfolder=kwargs.pop("""subfolder""" , __lowerCamelCase ) , cache_dir=kwargs.pop("""cache_dir""" , __lowerCamelCase ) , force_download=kwargs.pop("""force_download""" , __lowerCamelCase ) , proxies=kwargs.pop("""proxies""" , __lowerCamelCase ) , resume_download=kwargs.pop("""resume_download""" , __lowerCamelCase ) , local_files_only=kwargs.pop("""local_files_only""" , __lowerCamelCase ) , use_auth_token=kwargs.pop("""use_auth_token""" , __lowerCamelCase ) , revision=kwargs.pop("""revision""" , __lowerCamelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(__lowerCamelCase , __lowerCamelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) UpperCAmelCase = None else: with open(__lowerCamelCase ) as speaker_embeddings_json: UpperCAmelCase = json.load(__lowerCamelCase ) else: UpperCAmelCase = None UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) return cls(tokenizer=__lowerCamelCase , speaker_embeddings=__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int="speaker_embeddings_path.json" , __lowerCamelCase : Any="speaker_embeddings" , __lowerCamelCase : bool = False , **__lowerCamelCase : Any , ) -> str: """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(__lowerCamelCase , __lowerCamelCase , """v2""" ) , exist_ok=__lowerCamelCase ) UpperCAmelCase = {} UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": UpperCAmelCase = self._load_voice_preset(__lowerCamelCase ) UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["""repo_or_path"""] , __lowerCamelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=__lowerCamelCase , ) UpperCAmelCase = os.path.join(__lowerCamelCase , F"""{prompt_key}_{key}.npy""" ) UpperCAmelCase = tmp_dict with open(os.path.join(__lowerCamelCase , __lowerCamelCase ) , """w""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) super().save_pretrained(__lowerCamelCase , __lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : int , __lowerCamelCase : str = None , **__lowerCamelCase : Tuple ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.speaker_embeddings[voice_preset] UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get("""repo_or_path""" , """/""" ) , voice_preset_paths[key] , subfolder=kwargs.pop("""subfolder""" , __lowerCamelCase ) , cache_dir=kwargs.pop("""cache_dir""" , __lowerCamelCase ) , force_download=kwargs.pop("""force_download""" , __lowerCamelCase ) , proxies=kwargs.pop("""proxies""" , __lowerCamelCase ) , resume_download=kwargs.pop("""resume_download""" , __lowerCamelCase ) , local_files_only=kwargs.pop("""local_files_only""" , __lowerCamelCase ) , use_auth_token=kwargs.pop("""use_auth_token""" , __lowerCamelCase ) , revision=kwargs.pop("""revision""" , __lowerCamelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) UpperCAmelCase = np.load(__lowerCamelCase ) return voice_preset_dict def _lowercase ( self : Any , __lowerCamelCase : Optional[dict] = None ) -> Any: """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self : int , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=None , __lowerCamelCase : str="pt" , __lowerCamelCase : Any=2_5_6 , __lowerCamelCase : List[str]=False , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[Any]=False , **__lowerCamelCase : Optional[int] , ) -> List[Any]: """simple docstring""" if voice_preset is not None and not isinstance(__lowerCamelCase , __lowerCamelCase ): if ( isinstance(__lowerCamelCase , __lowerCamelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): UpperCAmelCase = self._load_voice_preset(__lowerCamelCase ) else: if isinstance(__lowerCamelCase , __lowerCamelCase ) and not voice_preset.endswith(""".npz""" ): UpperCAmelCase = voice_preset + """.npz""" UpperCAmelCase = np.load(__lowerCamelCase ) if voice_preset is not None: self._validate_voice_preset_dict(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = BatchFeature(data=__lowerCamelCase , tensor_type=__lowerCamelCase ) UpperCAmelCase = self.tokenizer( __lowerCamelCase , return_tensors=__lowerCamelCase , padding="""max_length""" , max_length=__lowerCamelCase , return_attention_mask=__lowerCamelCase , return_token_type_ids=__lowerCamelCase , add_special_tokens=__lowerCamelCase , **__lowerCamelCase , ) if voice_preset is not None: UpperCAmelCase = voice_preset return encoded_text
707
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
708
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
0
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer UpperCAmelCase = flax_key_tuple[:-1] + ("""weight""",) UpperCAmelCase = torch.permute(lowerCAmelCase_ , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(lowerCAmelCase_ ): # linear layer UpperCAmelCase = flax_key_tuple[:-1] + ("""weight""",) UpperCAmelCase = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: UpperCAmelCase = flax_key_tuple[:-1] + ("""weight""",) return flax_key_tuple, flax_tensor def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): if "metadata" in layer: UpperCAmelCase = layer.split("""metadata""" ) UpperCAmelCase = """""".join(split_layer[0] )[:-1] UpperCAmelCase = [tuple(("""metadata""" + split_layer[1]).split("""/""" ) )] elif "kvstore" in layer: UpperCAmelCase = layer.split("""kvstore""" ) UpperCAmelCase = """""".join(split_layer[0] )[:-1] UpperCAmelCase = [tuple(("""kvstore""" + split_layer[1]).split("""/""" ) )] else: UpperCAmelCase = layer.split("""/""" ) UpperCAmelCase = """/""".join(split_layer[:-1] ) UpperCAmelCase = (split_layer[-1],) if "kvstore/path" in layer: UpperCAmelCase = F"""{switch_checkpoint_path}/{checkpoint_info[layer]}""" elif "kvstore/driver" in layer: UpperCAmelCase = """file""" else: UpperCAmelCase = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = rename_keys(lowerCAmelCase_ ) UpperCAmelCase = {} for k, v in current_block.items(): UpperCAmelCase = v UpperCAmelCase = new_current_block torch.save(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = WEIGHTS_NAME ): UpperCAmelCase = convert_file_size_to_int(lowerCAmelCase_ ) UpperCAmelCase = [] UpperCAmelCase = {} UpperCAmelCase = 0 UpperCAmelCase = 0 os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) with gfile.GFile(switch_checkpoint_path + """/checkpoint""" , """rb""" ) as fp: UpperCAmelCase = serialization.msgpack_restore(fp.read() )["""optimizer"""]["""target"""] UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep="""/""" ) UpperCAmelCase = {} for layer in checkpoint_info.keys(): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = get_key_and_tensorstore_dict( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if curr_real_layer_name in all_layers: UpperCAmelCase = content else: UpperCAmelCase = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file UpperCAmelCase = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() UpperCAmelCase = torch.tensor(lowerCAmelCase_ ) UpperCAmelCase = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts UpperCAmelCase , UpperCAmelCase = rename_base_flax_keys(tuple(key.split("""/""" ) ) , lowerCAmelCase_ ) UpperCAmelCase = """/""".join(lowerCAmelCase_ ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: UpperCAmelCase = os.path.join( lowerCAmelCase_ , weights_name.replace(""".bin""" , F"""-{len(lowerCAmelCase_ )+1:05d}-of-???.bin""" ) ) rename_and_save_block(lowerCAmelCase_ , lowerCAmelCase_ ) sharded_state_dicts.append(current_block.keys() ) del current_block UpperCAmelCase = {} UpperCAmelCase = 0 UpperCAmelCase = raw_weights.to(getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) current_block_size += weight_size total_size += weight_size # Add the last block UpperCAmelCase = os.path.join(lowerCAmelCase_ , weights_name.replace(""".bin""" , F"""-{len(lowerCAmelCase_ )+1:05d}-of-???.bin""" ) ) rename_and_save_block(lowerCAmelCase_ , lowerCAmelCase_ ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(lowerCAmelCase_ ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index UpperCAmelCase = {} UpperCAmelCase = {} for idx, shard in enumerate(lowerCAmelCase_ ): UpperCAmelCase = weights_name.replace( """.bin""" , F"""-{idx+1:05d}-of-{len(lowerCAmelCase_ ):05d}.bin""" ) # len(sharded_state_dicts):05d} UpperCAmelCase = os.path.join(lowerCAmelCase_ , weights_name.replace(""".bin""" , F"""-{idx+1:05d}-of-???.bin""" ) ) os.rename(lowerCAmelCase_ , os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) ) UpperCAmelCase = shard for key in shard: UpperCAmelCase = shard_file # Add the metadata UpperCAmelCase = {"""total_size""": total_size} UpperCAmelCase = {"""metadata""": metadata, """weight_map""": weight_map} with open(os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) , """w""" , encoding="""utf-8""" ) as f: UpperCAmelCase = json.dumps(lowerCAmelCase_ , indent=2 , sort_keys=lowerCAmelCase_ ) + """\n""" f.write(lowerCAmelCase_ ) return metadata, index if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--max_shard_size""", default="""10GB""", required=False, help="""Max shard size""") parser.add_argument("""--dtype""", default="""bfloat16""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted""", type=str, required=False, help="""Path to the output pytorch model.""", ) __a = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def _UpperCamelCase ( ): from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer UpperCAmelCase = SwitchTransformersConfig.from_pretrained("""google/switch-base-8""" ) config.save_pretrained("""/home/arthur_huggingface_co/transformers/switch_converted""" ) UpperCAmelCase = SwitchTransformersForConditionalGeneration.from_pretrained( """/home/arthur_huggingface_co/transformers/switch_converted""" , device_map="""auto""" ) UpperCAmelCase = TaTokenizer.from_pretrained("""t5-small""" ) UpperCAmelCase = """A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.""" UpperCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors="""pt""" ).input_ids UpperCAmelCase = model.generate(lowerCAmelCase_ , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
709
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
0
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : Union[str, Any]=1_3 , __lowerCamelCase : Tuple=7 , __lowerCamelCase : int=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : Optional[Any]=9_9 , __lowerCamelCase : List[Any]=3_2 , __lowerCamelCase : Any=5 , __lowerCamelCase : Dict=4 , __lowerCamelCase : str=3_7 , __lowerCamelCase : Dict="gelu" , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : int=5_1_2 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : int=0.02 , __lowerCamelCase : Any=4 , ) -> Dict: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_attention_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_choices def _lowercase ( self : int ) -> str: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_attention_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = FlaxAlbertModelTester(self ) @slow def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase = model_class_name.from_pretrained("""albert-base-v2""" ) UpperCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowerCamelCase ) @require_flax class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = FlaxAlbertModel.from_pretrained("""albert-base-v2""" ) UpperCAmelCase = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) UpperCAmelCase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] UpperCAmelCase = (1, 1_1, 7_6_8) self.assertEqual(output.shape , __lowerCamelCase ) UpperCAmelCase = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , __lowerCamelCase , atol=1e-4 ) )
710
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
0
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = DanceDiffusionPipeline UpperCamelCase = UNCONDITIONAL_AUDIO_GENERATION_PARAMS UpperCamelCase = PipelineTesterMixin.required_optional_params - { '''callback''', '''latents''', '''callback_steps''', '''output_type''', '''num_images_per_prompt''', } UpperCamelCase = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : List[Any] ) -> int: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDModel( block_out_channels=(3_2, 3_2, 6_4) , extra_in_channels=1_6 , sample_size=5_1_2 , sample_rate=1_6_0_0_0 , in_channels=2 , out_channels=2 , flip_sin_to_cos=__lowerCamelCase , use_timestep_embedding=__lowerCamelCase , time_embedding_type="""fourier""" , mid_block_type="""UNetMidBlock1D""" , down_block_types=("""DownBlock1DNoSkip""", """DownBlock1D""", """AttnDownBlock1D""") , up_block_types=("""AttnUpBlock1D""", """UpBlock1D""", """UpBlock1DNoSkip""") , ) UpperCAmelCase = IPNDMScheduler() UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, } return components def _lowercase ( self : Any , __lowerCamelCase : List[str] , __lowerCamelCase : Optional[Any]=0 ) -> Union[str, Any]: """simple docstring""" if str(__lowerCamelCase ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(__lowerCamelCase ) else: UpperCAmelCase = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) UpperCAmelCase = { """batch_size""": 1, """generator""": generator, """num_inference_steps""": 4, } return inputs def _lowercase ( self : str ) -> List[str]: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = DanceDiffusionPipeline(**__lowerCamelCase ) UpperCAmelCase = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = self.get_dummy_inputs(__lowerCamelCase ) UpperCAmelCase = pipe(**__lowerCamelCase ) UpperCAmelCase = output.audios UpperCAmelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) UpperCAmelCase = np.array([-0.7_265, 1.0_000, -0.8_388, 0.1_175, 0.9_498, -1.0_000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def _lowercase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" return super().test_save_load_local() @skip_mps def _lowercase ( self : str ) -> Any: """simple docstring""" return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) @skip_mps def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" return super().test_save_load_optional_components() @skip_mps def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" return super().test_attention_slicing_forward_pass() def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = torch_device UpperCAmelCase = DanceDiffusionPipeline.from_pretrained("""harmonai/maestro-150k""" ) UpperCAmelCase = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = pipe(generator=__lowerCamelCase , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) UpperCAmelCase = output.audios UpperCAmelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCAmelCase = np.array([-0.0_192, -0.0_231, -0.0_318, -0.0_059, 0.0_002, -0.0_020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch_device UpperCAmelCase = DanceDiffusionPipeline.from_pretrained("""harmonai/maestro-150k""" , torch_dtype=torch.floataa ) UpperCAmelCase = pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = pipe(generator=__lowerCamelCase , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) UpperCAmelCase = output.audios UpperCAmelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCAmelCase = np.array([-0.0_367, -0.0_488, -0.0_771, -0.0_525, -0.0_444, -0.0_341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2
711
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class __lowercase : def __init__( self : str , __lowerCamelCase : Tuple , __lowerCamelCase : Any=1_3 , __lowerCamelCase : Any=7 , __lowerCamelCase : Dict=True , __lowerCamelCase : Any=True , __lowerCamelCase : Any=True , __lowerCamelCase : int=9_9 , __lowerCamelCase : Tuple=3_2 , __lowerCamelCase : List[str]=5 , __lowerCamelCase : int=4 , __lowerCamelCase : Optional[Any]=3_7 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Optional[Any]=5_1_2 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Optional[Any]=None , ) -> Dict: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope UpperCAmelCase = self.vocab_size - 1 def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) UpperCAmelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def _lowercase ( self : int , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Dict , __lowerCamelCase : Any , __lowerCamelCase : Tuple , *__lowerCamelCase : int ) -> List[Any]: """simple docstring""" UpperCAmelCase = OpenAIGPTModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , head_mask=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Dict , *__lowerCamelCase : Any ) -> Dict: """simple docstring""" UpperCAmelCase = OpenAIGPTLMHeadModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : int , __lowerCamelCase : List[Any] , __lowerCamelCase : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , *__lowerCamelCase : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = OpenAIGPTDoubleHeadsModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : Tuple , __lowerCamelCase : str , __lowerCamelCase : int , *__lowerCamelCase : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = OpenAIGPTForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : Dict ) -> int: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask, } return config, inputs_dict @require_torch class __lowercase ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) UpperCamelCase = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly UpperCamelCase = ( { '''feature-extraction''': OpenAIGPTModel, '''text-classification''': OpenAIGPTForSequenceClassification, '''text-generation''': OpenAIGPTLMHeadModel, '''zero-shot''': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def _lowercase ( self : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def _lowercase ( self : Tuple , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : List[str]=False ) -> List[Any]: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=__lowerCamelCase , ) UpperCAmelCase = inputs_dict["""labels"""] UpperCAmelCase = inputs_dict["""labels"""] UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=__lowerCamelCase , ) UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) return inputs_dict def _lowercase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = OpenAIGPTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , n_embd=3_7 ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : List[Any] ) -> int: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*__lowerCamelCase ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> str: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*__lowerCamelCase ) @slow def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = OpenAIGPTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @require_torch class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : Any ) -> Optional[int]: """simple docstring""" UpperCAmelCase = OpenAIGPTLMHeadModel.from_pretrained("""openai-gpt""" ) model.to(__lowerCamelCase ) UpperCAmelCase = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]] , dtype=torch.long , device=__lowerCamelCase ) # the president is UpperCAmelCase = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the UpperCAmelCase = model.generate(__lowerCamelCase , do_sample=__lowerCamelCase ) self.assertListEqual(output_ids[0].tolist() , __lowerCamelCase )
712
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
713
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
0
from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
714
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
0
def _UpperCamelCase ( lowerCAmelCase_ ) ->Dict: if not head: return True # split the list to two parts UpperCAmelCase , UpperCAmelCase = head.next, head while fast and fast.next: UpperCAmelCase = fast.next.next UpperCAmelCase = slow.next UpperCAmelCase = slow.next UpperCAmelCase = None # Don't forget here! But forget still works! # reverse the second part UpperCAmelCase = None while second: UpperCAmelCase = second.next UpperCAmelCase = node UpperCAmelCase = second UpperCAmelCase = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False UpperCAmelCase = node.next UpperCAmelCase = head.next return True def _UpperCamelCase ( lowerCAmelCase_ ) ->str: if not head or not head.next: return True # 1. Get the midpoint (slow) UpperCAmelCase = UpperCAmelCase = UpperCAmelCase = head while fast and fast.next: UpperCAmelCase , UpperCAmelCase = fast.next.next, slow.next # 2. Push the second half into the stack UpperCAmelCase = [slow.val] while slow.next: UpperCAmelCase = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False UpperCAmelCase = cur.next return True def _UpperCamelCase ( lowerCAmelCase_ ) ->Union[str, Any]: if not head or not head.next: return True UpperCAmelCase = {} UpperCAmelCase = 0 while head: if head.val in d: d[head.val].append(lowerCAmelCase_ ) else: UpperCAmelCase = [pos] UpperCAmelCase = head.next pos += 1 UpperCAmelCase = pos - 1 UpperCAmelCase = 0 for v in d.values(): if len(lowerCAmelCase_ ) % 2 != 0: middle += 1 else: UpperCAmelCase = 0 for i in range(0 , len(lowerCAmelCase_ ) ): if v[i] + v[len(lowerCAmelCase_ ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
715
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process __a = logging.getLogger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: return (preds == labels).mean() @dataclass class A : UpperCamelCase = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) UpperCamelCase = field( default=__snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) UpperCamelCase = field( default=__snake_case , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) UpperCamelCase = field( default=__snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class A : UpperCamelCase = field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(processors.keys() )} ) UpperCamelCase = field(metadata={'''help''': '''Should contain the data files for the task.'''} ) UpperCamelCase = field( default=1_28 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) UpperCamelCase = field( default=__snake_case , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def _UpperCamelCase ( ) ->Tuple: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , lowerCAmelCase_ ) # Set seed set_seed(training_args.seed ) try: UpperCAmelCase = processors[data_args.task_name]() UpperCAmelCase = processor.get_labels() UpperCAmelCase = len(lowerCAmelCase_ ) except KeyError: raise ValueError("""Task not found: %s""" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCAmelCase_ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) UpperCAmelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) UpperCAmelCase = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , ) # Get datasets UpperCAmelCase = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=lowerCAmelCase_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) UpperCAmelCase = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=lowerCAmelCase_ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(lowerCAmelCase_ ) -> Dict: UpperCAmelCase = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(lowerCAmelCase_ , p.label_ids )} # Data collator UpperCAmelCase = DataCollatorWithPadding(lowerCAmelCase_ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer UpperCAmelCase = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=lowerCAmelCase_ , eval_dataset=lowerCAmelCase_ , compute_metrics=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCAmelCase = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) UpperCAmelCase = trainer.evaluate() UpperCAmelCase = os.path.join(training_args.output_dir , """eval_results.txt""" ) if trainer.is_world_master(): with open(lowerCAmelCase_ , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""" , lowerCAmelCase_ , lowerCAmelCase_ ) writer.write("""%s = %s\n""" % (key, value) ) results.update(lowerCAmelCase_ ) return results def _UpperCamelCase ( lowerCAmelCase_ ) ->str: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
716
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class __lowercase ( __snake_case ): def __init__( self : Any , __lowerCamelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = data def __iter__( self : Optional[int] ) -> Optional[int]: """simple docstring""" for element in self.data: yield element def _UpperCamelCase ( lowerCAmelCase_=True ) ->List[Any]: UpperCAmelCase = Accelerator(even_batches=lowerCAmelCase_ ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = False ) ->Optional[Any]: if iterable: UpperCAmelCase = DummyIterableDataset(torch.as_tensor(range(lowerCAmelCase_ ) ) ) else: UpperCAmelCase = TensorDataset(torch.as_tensor(range(lowerCAmelCase_ ) ) ) UpperCAmelCase = DataLoader(lowerCAmelCase_ , batch_size=lowerCAmelCase_ ) UpperCAmelCase = accelerator.prepare(lowerCAmelCase_ ) return dl def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->Union[str, Any]: UpperCAmelCase = create_dataloader(accelerator=lowerCAmelCase_ , dataset_size=lowerCAmelCase_ , batch_size=lowerCAmelCase_ ) UpperCAmelCase = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def _UpperCamelCase ( ) ->str: UpperCAmelCase = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def _UpperCamelCase ( ) ->str: UpperCAmelCase = create_accelerator(even_batches=lowerCAmelCase_ ) verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( lowerCAmelCase_ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def _UpperCamelCase ( ) ->Optional[int]: UpperCAmelCase = create_accelerator(even_batches=lowerCAmelCase_ ) UpperCAmelCase = torch.nn.Linear(1 , 1 ) UpperCAmelCase = accelerator.prepare(lowerCAmelCase_ ) UpperCAmelCase = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) UpperCAmelCase = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ddp_model(batch[0].float() ) UpperCAmelCase = output.sum() loss.backward() batch_idxs.append(lowerCAmelCase_ ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def _UpperCamelCase ( lowerCAmelCase_ ) ->List[Any]: with warnings.catch_warnings(record=lowerCAmelCase_ ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , lowerCAmelCase_ ) assert "only supported for multi-GPU" in str(w[-1].message ) def _UpperCamelCase ( ) ->Optional[Any]: UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = create_accelerator(even_batches=lowerCAmelCase_ ) UpperCAmelCase = torch.nn.Linear(1 , 1 ) UpperCAmelCase = accelerator.prepare(lowerCAmelCase_ ) UpperCAmelCase = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) UpperCAmelCase = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase_ ): UpperCAmelCase = train_dl.batch_sampler.even_batches UpperCAmelCase = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def _UpperCamelCase ( ) ->Optional[int]: UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = create_accelerator(even_batches=lowerCAmelCase_ ) UpperCAmelCase = torch.nn.Linear(1 , 1 ) UpperCAmelCase = accelerator.prepare(lowerCAmelCase_ ) create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 , iterable=lowerCAmelCase_ ) UpperCAmelCase = create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings("""ignore""" ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase_ ): UpperCAmelCase = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def _UpperCamelCase ( ) ->List[str]: UpperCAmelCase = create_accelerator() UpperCAmelCase = torch.nn.Linear(1 , 1 ) UpperCAmelCase = accelerator.prepare(lowerCAmelCase_ ) create_dataloader(lowerCAmelCase_ , dataset_size=3 , batch_size=1 , iterable=lowerCAmelCase_ ) with warnings.catch_warnings(record=lowerCAmelCase_ ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowerCAmelCase_ ): pass assert issubclass(w[-1].category , lowerCAmelCase_ ) assert "only supported for map-style datasets" in str(w[-1].message ) def _UpperCamelCase ( ) ->str: UpperCAmelCase = create_accelerator() accelerator.print("""Test that even_batches variable ensures uniform batches across processes""" ) test_default_ensures_even_batch_sizes() accelerator.print("""Run tests with even_batches disabled""" ) test_can_disable_even_batches() accelerator.print("""Test joining uneven inputs""" ) test_can_join_uneven_inputs() accelerator.print("""Test overriding even_batches when joining uneven inputs""" ) test_join_can_override_even_batches() accelerator.print("""Test overriding even_batches for mixed dataloader types""" ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print("""Test overriding even_batches raises a warning for iterable dataloaders""" ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print("""Test join with non DDP distributed raises warning""" ) UpperCAmelCase = accelerator.state.distributed_type UpperCAmelCase = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(lowerCAmelCase_ ) UpperCAmelCase = original_state if __name__ == "__main__": main()
717
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/timesformer""": """https://huggingface.co/facebook/timesformer/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''timesformer''' def __init__( self : Union[str, Any] , __lowerCamelCase : Any=2_2_4 , __lowerCamelCase : Optional[Any]=1_6 , __lowerCamelCase : int=3 , __lowerCamelCase : Dict=8 , __lowerCamelCase : Dict=7_6_8 , __lowerCamelCase : int=1_2 , __lowerCamelCase : Any=1_2 , __lowerCamelCase : Dict=3_0_7_2 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Any=0.0 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : Dict=0.02 , __lowerCamelCase : int=1e-6 , __lowerCamelCase : Tuple=True , __lowerCamelCase : List[Any]="divided_space_time" , __lowerCamelCase : str=0 , **__lowerCamelCase : str , ) -> str: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = num_frames UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = qkv_bias UpperCAmelCase = attention_type UpperCAmelCase = drop_path_rate
718
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
0
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->float | int: for nxt, d in graph[v]: if nxt in visited_forward: continue UpperCAmelCase = cst_fwd.get(lowerCAmelCase_ , np.inf ) UpperCAmelCase = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) UpperCAmelCase = new_cost_f UpperCAmelCase = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: UpperCAmelCase = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->int: UpperCAmelCase = -1 UpperCAmelCase = set() UpperCAmelCase = set() UpperCAmelCase = {source: 0} UpperCAmelCase = {destination: 0} UpperCAmelCase = {source: None} UpperCAmelCase = {destination: None} UpperCAmelCase = PriorityQueue() UpperCAmelCase = PriorityQueue() UpperCAmelCase = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): UpperCAmelCase , UpperCAmelCase = queue_forward.get() visited_forward.add(lowerCAmelCase_ ) UpperCAmelCase , UpperCAmelCase = queue_backward.get() visited_backward.add(lowerCAmelCase_ ) UpperCAmelCase = pass_and_relaxation( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) UpperCAmelCase = pass_and_relaxation( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: UpperCAmelCase = shortest_distance return shortest_path_distance __a = { """B""": [["""C""", 1]], """C""": [["""D""", 1]], """D""": [["""F""", 1]], """E""": [["""B""", 1], ["""G""", 2]], """F""": [], """G""": [["""F""", 1]], } __a = { """B""": [["""E""", 1]], """C""": [["""B""", 1]], """D""": [["""C""", 1]], """F""": [["""D""", 1], ["""G""", 1]], """E""": [[None, np.inf]], """G""": [["""E""", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
719
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """bert-base-uncased""": """https://huggingface.co/bert-base-uncased/resolve/main/config.json""", """bert-large-uncased""": """https://huggingface.co/bert-large-uncased/resolve/main/config.json""", """bert-base-cased""": """https://huggingface.co/bert-base-cased/resolve/main/config.json""", """bert-large-cased""": """https://huggingface.co/bert-large-cased/resolve/main/config.json""", """bert-base-multilingual-uncased""": """https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json""", """bert-base-multilingual-cased""": """https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json""", """bert-base-chinese""": """https://huggingface.co/bert-base-chinese/resolve/main/config.json""", """bert-base-german-cased""": """https://huggingface.co/bert-base-german-cased/resolve/main/config.json""", """bert-large-uncased-whole-word-masking""": ( """https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json""" ), """bert-large-cased-whole-word-masking""": ( """https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json""" ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( """https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json""" ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( """https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json""" ), """bert-base-cased-finetuned-mrpc""": """https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json""", """bert-base-german-dbmdz-cased""": """https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json""", """bert-base-german-dbmdz-uncased""": """https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json""", """cl-tohoku/bert-base-japanese""": """https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json""", """cl-tohoku/bert-base-japanese-whole-word-masking""": ( """https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json""" ), """cl-tohoku/bert-base-japanese-char""": ( """https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json""" ), """cl-tohoku/bert-base-japanese-char-whole-word-masking""": ( """https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json""" ), """TurkuNLP/bert-base-finnish-cased-v1""": ( """https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json""" ), """TurkuNLP/bert-base-finnish-uncased-v1""": ( """https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json""" ), """wietsedv/bert-base-dutch-cased""": """https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json""", # See all BERT models at https://huggingface.co/models?filter=bert } class __lowercase ( __snake_case ): UpperCamelCase = '''bert''' def __init__( self : List[str] , __lowerCamelCase : Tuple=3_0_5_2_2 , __lowerCamelCase : Optional[int]=7_6_8 , __lowerCamelCase : Tuple=1_2 , __lowerCamelCase : str=1_2 , __lowerCamelCase : Tuple=3_0_7_2 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : List[str]=5_1_2 , __lowerCamelCase : int=2 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : str=1e-1_2 , __lowerCamelCase : Any=0 , __lowerCamelCase : Any="absolute" , __lowerCamelCase : int=True , __lowerCamelCase : Any=None , **__lowerCamelCase : List[Any] , ) -> Dict: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = position_embedding_type UpperCAmelCase = use_cache UpperCAmelCase = classifier_dropout class __lowercase ( __snake_case ): @property def _lowercase ( self : int ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
720
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
0
'''simple docstring''' import comet # From: unbabel-comet import torch import datasets __a = datasets.logging.get_logger(__name__) __a = """\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel's Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = \"{COMET}: A Neural Framework for {MT} Evaluation\", author = \"Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon\", booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\", month = nov, year = \"2020\", address = \"Online\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\", pages = \"2685--2702\", } """ __a = """\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. """ __a = """ COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric('comet') >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"] >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"] >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results[\"scores\"]]) [0.19, 0.92] """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://unbabel.github.io/COMET/html/index.html""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """sources""": datasets.Value("""string""" , id="""sequence""" ), """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/Unbabel/COMET"""] , reference_urls=[ """https://github.com/Unbabel/COMET""", """https://www.aclweb.org/anthology/2020.emnlp-main.213/""", """http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6""", ] , ) def _lowercase ( self : str , __lowerCamelCase : Union[str, Any] ) -> Optional[Any]: """simple docstring""" if self.config_name == "default": UpperCAmelCase = comet.load_from_checkpoint(comet.download_model("""wmt20-comet-da""" ) ) else: UpperCAmelCase = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def _lowercase ( self : str , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : int=False ) -> List[Any]: """simple docstring""" if gpus is None: UpperCAmelCase = 1 if torch.cuda.is_available() else 0 UpperCAmelCase = {"""src""": sources, """mt""": predictions, """ref""": references} UpperCAmelCase = [dict(zip(__lowerCamelCase , __lowerCamelCase ) ) for t in zip(*data.values() )] UpperCAmelCase , UpperCAmelCase = self.scorer.predict(__lowerCamelCase , gpus=__lowerCamelCase , progress_bar=__lowerCamelCase ) return {"mean_score": mean_score, "scores": scores}
721
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
0
from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP __a = logging.get_logger(__name__) # pylint: disable=invalid-name __a = """ Examples: ```py >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline >>> import torch >>> pipe_prior = KandinskyPriorPipeline.from_pretrained(\"kandinsky-community/Kandinsky-2-1-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> negative_image_emb = out.negative_image_embeds >>> pipe = KandinskyPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-1\") >>> pipe.to(\"cuda\") >>> image = pipe( ... prompt, ... image_embeds=image_emb, ... negative_image_embeds=negative_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=8 ) ->List[str]: UpperCAmelCase = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 UpperCAmelCase = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class __lowercase ( __snake_case ): def __init__( self : Optional[Any] , __lowerCamelCase : MultilingualCLIP , __lowerCamelCase : XLMRobertaTokenizer , __lowerCamelCase : UNetaDConditionModel , __lowerCamelCase : Union[DDIMScheduler, DDPMScheduler] , __lowerCamelCase : VQModel , ) -> List[str]: """simple docstring""" super().__init__() self.register_modules( text_encoder=__lowerCamelCase , tokenizer=__lowerCamelCase , unet=__lowerCamelCase , scheduler=__lowerCamelCase , movq=__lowerCamelCase , ) UpperCAmelCase = 2 ** (len(self.movq.config.block_out_channels ) - 1) def _lowercase ( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : Any ) -> Tuple: """simple docstring""" if latents is None: UpperCAmelCase = randn_tensor(__lowerCamelCase , generator=__lowerCamelCase , device=__lowerCamelCase , dtype=__lowerCamelCase ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase = latents.to(__lowerCamelCase ) UpperCAmelCase = latents * scheduler.init_noise_sigma return latents def _lowercase ( self : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any]=None , ) -> List[str]: """simple docstring""" UpperCAmelCase = len(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else 1 # get prompt text embeddings UpperCAmelCase = self.tokenizer( __lowerCamelCase , padding="""max_length""" , truncation=__lowerCamelCase , max_length=7_7 , return_attention_mask=__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_tensors="""pt""" , ) UpperCAmelCase = text_inputs.input_ids UpperCAmelCase = self.tokenizer(__lowerCamelCase , padding="""longest""" , return_tensors="""pt""" ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" F""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) UpperCAmelCase = text_input_ids.to(__lowerCamelCase ) UpperCAmelCase = text_inputs.attention_mask.to(__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = self.text_encoder( input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ) UpperCAmelCase = prompt_embeds.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = text_encoder_hidden_states.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = text_mask.repeat_interleave(__lowerCamelCase , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase = 4_2 if negative_prompt is None: UpperCAmelCase = [""""""] * batch_size elif type(__lowerCamelCase ) is not type(__lowerCamelCase ): raise TypeError( F"""`negative_prompt` should be the same type to `prompt`, but got {type(__lowerCamelCase )} !=""" F""" {type(__lowerCamelCase )}.""" ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = [negative_prompt] elif batch_size != len(__lowerCamelCase ): raise ValueError( F"""`negative_prompt`: {negative_prompt} has batch size {len(__lowerCamelCase )}, but `prompt`:""" F""" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches""" """ the batch size of `prompt`.""" ) else: UpperCAmelCase = negative_prompt UpperCAmelCase = self.tokenizer( __lowerCamelCase , padding="""max_length""" , max_length=7_7 , truncation=__lowerCamelCase , return_attention_mask=__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_tensors="""pt""" , ) UpperCAmelCase = uncond_input.input_ids.to(__lowerCamelCase ) UpperCAmelCase = uncond_input.attention_mask.to(__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = self.text_encoder( input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method UpperCAmelCase = negative_prompt_embeds.shape[1] UpperCAmelCase = negative_prompt_embeds.repeat(1 , __lowerCamelCase ) UpperCAmelCase = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __lowerCamelCase ) UpperCAmelCase = uncond_text_encoder_hidden_states.shape[1] UpperCAmelCase = uncond_text_encoder_hidden_states.repeat(1 , __lowerCamelCase , 1 ) UpperCAmelCase = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , __lowerCamelCase , -1 ) UpperCAmelCase = uncond_text_mask.repeat_interleave(__lowerCamelCase , dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase = torch.cat([negative_prompt_embeds, prompt_embeds] ) UpperCAmelCase = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) UpperCAmelCase = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def _lowercase ( self : Tuple , __lowerCamelCase : Dict=0 ) -> Any: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) UpperCAmelCase = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any]=0 ) -> List[str]: """simple docstring""" if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) UpperCAmelCase = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=__lowerCamelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: UpperCAmelCase , UpperCAmelCase = cpu_offload_with_hook(__lowerCamelCase , __lowerCamelCase , prev_module_hook=__lowerCamelCase ) if self.safety_checker is not None: UpperCAmelCase , UpperCAmelCase = cpu_offload_with_hook(self.safety_checker , __lowerCamelCase , prev_module_hook=__lowerCamelCase ) # We'll offload the last model manually. UpperCAmelCase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(__lowerCamelCase , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(__lowerCamelCase ) def __call__( self : List[str] , __lowerCamelCase : Union[str, List[str]] , __lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , __lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , __lowerCamelCase : Optional[Union[str, List[str]]] = None , __lowerCamelCase : int = 5_1_2 , __lowerCamelCase : int = 5_1_2 , __lowerCamelCase : int = 1_0_0 , __lowerCamelCase : float = 4.0 , __lowerCamelCase : int = 1 , __lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __lowerCamelCase : Optional[torch.FloatTensor] = None , __lowerCamelCase : Optional[str] = "pil" , __lowerCamelCase : bool = True , ) -> Any: """simple docstring""" if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = 1 elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = len(__lowerCamelCase ) else: raise ValueError(F"""`prompt` has to be of type `str` or `list` but is {type(__lowerCamelCase )}""" ) UpperCAmelCase = self._execution_device UpperCAmelCase = batch_size * num_images_per_prompt UpperCAmelCase = guidance_scale > 1.0 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self._encode_prompt( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = torch.cat(__lowerCamelCase , dim=0 ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = torch.cat(__lowerCamelCase , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase = image_embeds.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = negative_image_embeds.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to( dtype=prompt_embeds.dtype , device=__lowerCamelCase ) self.scheduler.set_timesteps(__lowerCamelCase , device=__lowerCamelCase ) UpperCAmelCase = self.scheduler.timesteps UpperCAmelCase = self.unet.config.in_channels UpperCAmelCase , UpperCAmelCase = get_new_h_w(__lowerCamelCase , __lowerCamelCase , self.movq_scale_factor ) # create initial latent UpperCAmelCase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(__lowerCamelCase ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase = {"""text_embeds""": prompt_embeds, """image_embeds""": image_embeds} UpperCAmelCase = self.unet( sample=__lowerCamelCase , timestep=__lowerCamelCase , encoder_hidden_states=__lowerCamelCase , added_cond_kwargs=__lowerCamelCase , return_dict=__lowerCamelCase , )[0] if do_classifier_free_guidance: UpperCAmelCase , UpperCAmelCase = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase , UpperCAmelCase = noise_pred.chunk(2 ) UpperCAmelCase , UpperCAmelCase = variance_pred.chunk(2 ) UpperCAmelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase , UpperCAmelCase = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase = self.scheduler.step( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase , ).prev_sample # post-processing UpperCAmelCase = self.movq.decode(__lowerCamelCase , force_not_quantize=__lowerCamelCase )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase = image * 0.5 + 0.5 UpperCAmelCase = image.clamp(0 , 1 ) UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase = self.numpy_to_pil(__lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__lowerCamelCase )
700
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() __a = logging.get_logger(__name__) __a = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """adapter_layer""": """encoder.layers.*.adapter_layer""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", """pooling_layer.linear""": """projector""", """pooling_layer.projection""": """classifier""", } __a = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", """projector""", """classifier""", ] def _UpperCamelCase ( lowerCAmelCase_ ) ->Tuple: UpperCAmelCase = {} with open(lowerCAmelCase_ , """r""" ) as file: for line_number, line in enumerate(lowerCAmelCase_ ): UpperCAmelCase = line.strip() if line: UpperCAmelCase = line.split() UpperCAmelCase = line_number UpperCAmelCase = words[0] UpperCAmelCase = value return result def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Optional[int]: for attribute in key.split(""".""" ): UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowerCAmelCase_ ): UpperCAmelCase = PARAM_MAPPING[full_name.split(""".""" )[-1]] UpperCAmelCase = """param""" if weight_type is not None and weight_type != "param": UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ).shape elif weight_type is not None and weight_type == "param": UpperCAmelCase = hf_pointer for attribute in hf_param_name.split(""".""" ): UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = shape_pointer.shape # let's reduce dimension UpperCAmelCase = value[0] else: UpperCAmelCase = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase = value elif weight_type == "weight_g": UpperCAmelCase = value elif weight_type == "weight_v": UpperCAmelCase = value elif weight_type == "bias": UpperCAmelCase = value elif weight_type == "param": for attribute in hf_param_name.split(""".""" ): UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = value else: UpperCAmelCase = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Optional[int]: UpperCAmelCase = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(lowerCAmelCase_ ): UpperCAmelCase = PARAM_MAPPING[full_name.split(""".""" )[-1]] UpperCAmelCase = """param""" if weight_type is not None and weight_type != "param": UpperCAmelCase = """.""".join([key, weight_type] ) elif weight_type is not None and weight_type == "param": UpperCAmelCase = """.""".join([key, hf_param_name] ) else: UpperCAmelCase = key UpperCAmelCase = value if """lm_head""" in full_key else value[0] __a = { """W_a""": """linear_1.weight""", """W_b""": """linear_2.weight""", """b_a""": """linear_1.bias""", """b_b""": """linear_2.bias""", """ln_W""": """norm.weight""", """ln_b""": """norm.bias""", } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None ) ->Optional[Any]: UpperCAmelCase = False for key, mapped_key in MAPPING.items(): UpperCAmelCase = """wav2vec2.""" + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: UpperCAmelCase = True if "*" in mapped_key: UpperCAmelCase = name.split(lowerCAmelCase_ )[0].split(""".""" )[-2] UpperCAmelCase = mapped_key.replace("""*""" , lowerCAmelCase_ ) if "weight_g" in name: UpperCAmelCase = """weight_g""" elif "weight_v" in name: UpperCAmelCase = """weight_v""" elif "bias" in name: UpperCAmelCase = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase = """weight""" else: UpperCAmelCase = None if hf_dict is not None: rename_dict(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) else: set_recursively(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) return is_used return is_used def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: UpperCAmelCase = [] UpperCAmelCase = fairseq_model.state_dict() UpperCAmelCase = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase = False if "conv_layers" in name: load_conv_layer( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , hf_model.config.feat_extract_norm == """group""" , ) UpperCAmelCase = True else: UpperCAmelCase = load_wavaveca_layer(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if not is_used: unused_weights.append(lowerCAmelCase_ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: UpperCAmelCase = full_name.split("""conv_layers.""" )[-1] UpperCAmelCase = name.split(""".""" ) UpperCAmelCase = int(items[0] ) UpperCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowerCAmelCase_ ) @torch.no_grad() def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False ) ->str: if config_path is not None: UpperCAmelCase = WavaVecaConfig.from_pretrained(lowerCAmelCase_ ) else: UpperCAmelCase = WavaVecaConfig() if is_seq_class: UpperCAmelCase = read_txt_into_dict(lowerCAmelCase_ ) UpperCAmelCase = idalabel UpperCAmelCase = WavaVecaForSequenceClassification(lowerCAmelCase_ ) UpperCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ) feature_extractor.save_pretrained(lowerCAmelCase_ ) elif is_finetuned: if dict_path: UpperCAmelCase = Dictionary.load(lowerCAmelCase_ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase = target_dict.pad_index UpperCAmelCase = target_dict.bos_index UpperCAmelCase = target_dict.eos_index UpperCAmelCase = len(target_dict.symbols ) UpperCAmelCase = os.path.join(lowerCAmelCase_ , """vocab.json""" ) if not os.path.isdir(lowerCAmelCase_ ): logger.error("""--pytorch_dump_folder_path ({}) should be a directory""".format(lowerCAmelCase_ ) ) return os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) UpperCAmelCase = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase = 0 UpperCAmelCase = 1 with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as vocab_handle: json.dump(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = WavaVecaCTCTokenizer( lowerCAmelCase_ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="""|""" , do_lower_case=lowerCAmelCase_ , ) UpperCAmelCase = True if config.feat_extract_norm == """layer""" else False UpperCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ) UpperCAmelCase = WavaVecaProcessor(feature_extractor=lowerCAmelCase_ , tokenizer=lowerCAmelCase_ ) processor.save_pretrained(lowerCAmelCase_ ) UpperCAmelCase = WavaVecaForCTC(lowerCAmelCase_ ) else: UpperCAmelCase = WavaVecaForPreTraining(lowerCAmelCase_ ) if is_finetuned or is_seq_class: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) else: UpperCAmelCase = argparse.Namespace(task="""audio_pretraining""" ) UpperCAmelCase = fairseq.tasks.setup_task(lowerCAmelCase_ ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=lowerCAmelCase_ ) UpperCAmelCase = model[0].eval() recursively_load_weights(lowerCAmelCase_ , lowerCAmelCase_ , not is_finetuned ) hf_wavavec.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) parser.add_argument( """--is_seq_class""", action="""store_true""", help="""Whether the model to convert is a fine-tuned sequence classification model or not""", ) __a = parser.parse_args() __a = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
701
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
0
def _UpperCamelCase ( lowerCAmelCase_ = 2_0_0_0_0_0_0 ) ->int: UpperCAmelCase = [0 for i in range(n + 1 )] UpperCAmelCase = 1 UpperCAmelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowerCAmelCase_ ): UpperCAmelCase = 1 UpperCAmelCase = 0 for i in range(lowerCAmelCase_ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F"""{solution() = }""")
702
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
0
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse("""3.8"""): import importlib_metadata else: import importlib.metadata as importlib_metadata __a = """""" if version.parse(importlib_metadata.version("""jiwer""")) < version.parse("""2.3.0"""): class __lowercase ( tr.AbstractTransform ): def __init__( self : Any , __lowerCamelCase : str = " " ) -> List[Any]: """simple docstring""" UpperCAmelCase = sentence_delimiter def _lowercase ( self : List[Any] , __lowerCamelCase : str ) -> int: """simple docstring""" return list(__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : List[str] ) -> List[str]: """simple docstring""" UpperCAmelCase = [] for sent_idx, sentence in enumerate(__lowerCamelCase ): chars.extend(self.process_string(__lowerCamelCase ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(__lowerCamelCase ) - 1: chars.append(self.sentence_delimiter ) return chars __a = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: __a = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) __a = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ __a = """\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER's output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. """ __a = """ Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> cer = datasets.load_metric(\"cer\") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : Tuple ) -> int: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", """https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates""", ] , ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any]=False ) -> List[str]: """simple docstring""" if concatenate_texts: return jiwer.compute_measures( __lowerCamelCase , __lowerCamelCase , truth_transform=__lowerCamelCase , hypothesis_transform=__lowerCamelCase , )["wer"] UpperCAmelCase = 0 UpperCAmelCase = 0 for prediction, reference in zip(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = jiwer.compute_measures( __lowerCamelCase , __lowerCamelCase , truth_transform=__lowerCamelCase , hypothesis_transform=__lowerCamelCase , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
703
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
0
import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) __a = [ """cross_validation.py""", """gradient_accumulation.py""", """local_sgd.py""", """multi_process_metrics.py""", """memory.py""", """automatic_gradient_accumulation.py""", """fsdp_with_peak_mem_tracking.py""", """deepspeed_with_config_support.py""", """megatron_lm_gpt_pretraining.py""", ] class __lowercase ( unittest.TestCase ): def _lowercase ( self : Dict , __lowerCamelCase : str , __lowerCamelCase : bool , __lowerCamelCase : str = None , __lowerCamelCase : list = None ) -> List[Any]: """simple docstring""" UpperCAmelCase = None UpperCAmelCase = os.path.abspath(os.path.join("""examples""" , """by_feature""" ) ) UpperCAmelCase = os.path.abspath("""examples""" ) for item in os.listdir(__lowerCamelCase ): if item not in EXCLUDE_EXAMPLES: UpperCAmelCase = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.isfile(__lowerCamelCase ) and ".py" in item_path: with self.subTest( tested_script=__lowerCamelCase , feature_script=__lowerCamelCase , tested_section="""main()""" if parser_only else """training_function()""" , ): UpperCAmelCase = compare_against_test( os.path.join(__lowerCamelCase , __lowerCamelCase ) , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = """\n""".join(__lowerCamelCase ) if special_strings is not None: for string in special_strings: UpperCAmelCase = diff.replace(__lowerCamelCase , """""" ) self.assertEqual(__lowerCamelCase , """""" ) def _lowercase ( self : str ) -> Optional[int]: """simple docstring""" self.one_complete_example("""complete_nlp_example.py""" , __lowerCamelCase ) self.one_complete_example("""complete_nlp_example.py""" , __lowerCamelCase ) def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = os.path.abspath(os.path.join("""examples""" , """cv_example.py""" ) ) UpperCAmelCase = [ """ """ * 1_6 + """{\n\n""", """ """ * 2_0 + """\"accuracy\": eval_metric[\"accuracy\"],\n\n""", """ """ * 2_0 + """\"f1\": eval_metric[\"f1\"],\n\n""", """ """ * 2_0 + """\"train_loss\": total_loss.item() / len(train_dataloader),\n\n""", """ """ * 2_0 + """\"epoch\": epoch,\n\n""", """ """ * 1_6 + """},\n\n""", """ """ * 1_6 + """step=epoch,\n""", """ """ * 1_2, """ """ * 8 + """for step, batch in enumerate(active_dataloader):\n""", ] self.one_complete_example("""complete_cv_example.py""" , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) self.one_complete_example("""complete_cv_example.py""" , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) @mock.patch.dict(os.environ , {'''TESTING_MOCKED_DATALOADERS''': '''1'''} ) class __lowercase ( __snake_case ): UpperCamelCase = False @classmethod def _lowercase ( cls : Any ) -> List[str]: """simple docstring""" super().setUpClass() UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = os.path.join(cls._tmpdir , """default_config.yml""" ) write_basic_config(save_location=cls.configPath ) UpperCAmelCase = ["""accelerate""", """launch""", """--config_file""", cls.configPath] @classmethod def _lowercase ( cls : Dict ) -> Optional[int]: """simple docstring""" super().tearDownClass() shutil.rmtree(cls._tmpdir ) def _lowercase ( self : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase = F""" examples/by_feature/checkpointing.py --checkpointing_steps epoch --output_dir {self.tmpdir} """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , """epoch_0""" ) ) ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = F""" examples/by_feature/checkpointing.py --checkpointing_steps 1 --output_dir {self.tmpdir} """.split() UpperCAmelCase = run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , """step_2""" ) ) ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = F""" examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , 'epoch_0' )} """.split() UpperCAmelCase = run_command(self._launch_args + testargs , return_stdout=__lowerCamelCase ) self.assertNotIn("""epoch 0:""" , __lowerCamelCase ) self.assertIn("""epoch 1:""" , __lowerCamelCase ) def _lowercase ( self : int ) -> str: """simple docstring""" UpperCAmelCase = F""" examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , 'step_2' )} """.split() UpperCAmelCase = run_command(self._launch_args + testargs , return_stdout=__lowerCamelCase ) if torch.cuda.is_available(): UpperCAmelCase = torch.cuda.device_count() else: UpperCAmelCase = 1 if num_processes > 1: self.assertNotIn("""epoch 0:""" , __lowerCamelCase ) self.assertIn("""epoch 1:""" , __lowerCamelCase ) else: self.assertIn("""epoch 0:""" , __lowerCamelCase ) self.assertIn("""epoch 1:""" , __lowerCamelCase ) @slow def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """ examples/by_feature/cross_validation.py --num_folds 2 """.split() with mock.patch.dict(os.environ , {"""TESTING_MOCKED_DATALOADERS""": """0"""} ): UpperCAmelCase = run_command(self._launch_args + testargs , return_stdout=__lowerCamelCase ) UpperCAmelCase = re.findall("""({.+})""" , __lowerCamelCase ) UpperCAmelCase = [r for r in results if """accuracy""" in r][-1] UpperCAmelCase = ast.literal_eval(__lowerCamelCase ) self.assertGreaterEqual(results["""accuracy"""] , 0.75 ) def _lowercase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = ["""examples/by_feature/multi_process_metrics.py"""] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ , {"""WANDB_MODE""": """offline"""} ) def _lowercase ( self : Union[str, Any] ) -> Dict: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: UpperCAmelCase = F""" examples/by_feature/tracking.py --with_tracking --project_dir {tmpdir} """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(__lowerCamelCase , """tracking""" ) ) ) def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = ["""examples/by_feature/gradient_accumulation.py"""] run_command(self._launch_args + testargs ) def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = ["""examples/by_feature/local_sgd.py"""] run_command(self._launch_args + testargs )
704
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
0
'''simple docstring''' def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = len(lowerCAmelCase_ ) + 1 UpperCAmelCase = len(lowerCAmelCase_ ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. UpperCAmelCase = [[0 for i in range(lowerCAmelCase_ )] for j in range(lowerCAmelCase_ )] # since string of zero length match pattern of zero length UpperCAmelCase = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , lowerCAmelCase_ ): UpperCAmelCase = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , lowerCAmelCase_ ): UpperCAmelCase = dp[0][j - 2] if pattern[j - 1] == """*""" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , lowerCAmelCase_ ): for j in range(1 , lowerCAmelCase_ ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": UpperCAmelCase = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: UpperCAmelCase = 1 elif pattern[j - 2] in (input_string[i - 1], "."): UpperCAmelCase = dp[i - 1][j] else: UpperCAmelCase = 0 else: UpperCAmelCase = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") __a = """aab""" __a = """c*a*b""" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F"""{input_string} matches the given pattern {pattern}""") else: print(F"""{input_string} does not match with the given pattern {pattern}""")
705
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
0
'''simple docstring''' from math import factorial def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->float: if successes > trials: raise ValueError("""successes must be lower or equal to trials""" ) if trials < 0 or successes < 0: raise ValueError("""the function is defined for non-negative integers""" ) if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError("""the function is defined for non-negative integers""" ) if not 0 < prob < 1: raise ValueError("""prob has to be in range of 1 - 0""" ) UpperCAmelCase = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! UpperCAmelCase = float(factorial(lowerCAmelCase_ ) ) coefficient /= factorial(lowerCAmelCase_ ) * factorial(trials - successes ) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print("""Probability of 2 successes out of 4 trails""") print("""with probability of 0.75 is:""", end=""" """) print(binomial_distribution(2, 4, 0.75))
706
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = { """configuration_trajectory_transformer""": [ """TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrajectoryTransformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrajectoryTransformerModel""", """TrajectoryTransformerPreTrainedModel""", """load_tf_weights_in_trajectory_transformer""", ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
707
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
0
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase = FlaxDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=__lowerCamelCase , cache_dir=__lowerCamelCase ) UpperCAmelCase = [t[-1] for t in os.walk(os.path.join(__lowerCamelCase , os.listdir(__lowerCamelCase )[0] , """snapshots""" ) )] UpperCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(""".bin""" ) for f in files ) @slow @require_flax class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=__lowerCamelCase ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 4 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 6_4, 6_4, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_514_745 ) < 1e-3 assert np.abs(np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 4_9_9_4_7.8_7_5 ) < 5e-1 UpperCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__lowerCamelCase ) == num_samples def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""flax""" , safety_checker=__lowerCamelCase ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_652_401) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_3_8_3_8_0_8.2) ) < 5e-1 def _lowercase ( self : int ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCamelCase ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_003_906) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_3_7_3_5_1_6.7_5) ) < 5e-1 def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_003_906) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_3_7_3_5_1_6.7_5) ) < 5e-1 def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = FlaxDDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , set_alpha_to_one=__lowerCamelCase , steps_offset=1 , ) UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , scheduler=__lowerCamelCase , safety_checker=__lowerCamelCase , ) UpperCAmelCase = scheduler.create_state() UpperCAmelCase = scheduler_state UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.045_043_945) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_3_4_7_6_9_3.5) ) < 5e-1 def _lowercase ( self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = jax.random.split(jax.random.PRNGKey(0 ) , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCamelCase , ) UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) UpperCAmelCase = images[2, 0, 2_5_6, 1_0:1_7, 1] # With memory efficient attention UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCamelCase , use_memory_efficient_attention=__lowerCamelCase , ) UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images_eff.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) UpperCAmelCase = images[2, 0, 2_5_6, 1_0:1_7, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1e-2
708
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
0
from __future__ import annotations import requests def _UpperCamelCase ( lowerCAmelCase_ ): UpperCAmelCase = F"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def _UpperCamelCase ( lowerCAmelCase_ = 1_0 ): UpperCAmelCase = """https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty""" UpperCAmelCase = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def _UpperCamelCase ( lowerCAmelCase_ = 1_0 ): UpperCAmelCase = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join("""* [{title}]({url})""".format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
709
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
0
def _UpperCamelCase ( lowerCAmelCase_ ) ->list[list]: UpperCAmelCase = current_set.copy() for row_index, row in enumerate(lowerCAmelCase_ ): UpperCAmelCase = row[0] for column_index, column in enumerate(lowerCAmelCase_ ): if magnitude == 0: UpperCAmelCase = column continue UpperCAmelCase = column / magnitude # Subtract to cancel term UpperCAmelCase = current_set[0] UpperCAmelCase = [first_row] UpperCAmelCase = current_set[1::] for row in current_set: UpperCAmelCase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(lowerCAmelCase_ ) continue for column_index in range(len(lowerCAmelCase_ ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(lowerCAmelCase_ ) # Create next recursion iteration set if len(final_set[0] ) != 3: UpperCAmelCase = final_set[0] UpperCAmelCase = [] UpperCAmelCase = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) UpperCAmelCase = simplify(lowerCAmelCase_ ) for i in range(len(lowerCAmelCase_ ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , lowerCAmelCase_ ) UpperCAmelCase = resultant return final_set def _UpperCamelCase ( lowerCAmelCase_ ) ->list: if len(lowerCAmelCase_ ) == 0: raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) UpperCAmelCase = len(lowerCAmelCase_ ) + 1 if any(len(lowerCAmelCase_ ) != _length for item in equations ): raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) for row in equations: if any(not isinstance(lowerCAmelCase_ , (int, float) ) for column in row ): raise ValueError("""solve_simultaneous() requires lists of integers""" ) if len(lowerCAmelCase_ ) == 1: return [equations[0][-1] / equations[0][0]] UpperCAmelCase = equations.copy() if any(0 in row for row in data_set ): UpperCAmelCase = data_set.copy() UpperCAmelCase = [] for row_index, row in enumerate(lowerCAmelCase_ ): if 0 not in row: UpperCAmelCase = data_set.pop(lowerCAmelCase_ ) break if not full_row: raise ValueError("""solve_simultaneous() requires at least 1 full equation""" ) data_set.insert(0 , lowerCAmelCase_ ) UpperCAmelCase = data_set.copy() UpperCAmelCase = simplify(lowerCAmelCase_ ) UpperCAmelCase = simplified[::-1] UpperCAmelCase = [] for row in simplified: UpperCAmelCase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue UpperCAmelCase = row.copy()[: len(lowerCAmelCase_ ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(lowerCAmelCase_ ) == 0: solutions.append(0 ) continue UpperCAmelCase = temp_row[1::] UpperCAmelCase = temp_row[::-1] for column_index, column in enumerate(lowerCAmelCase_ ): current_solution -= column * solutions[column_index] solutions.append(lowerCAmelCase_ ) UpperCAmelCase = [] for item in solutions: final.append(float(round(lowerCAmelCase_ , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() __a = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
710
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = BertJapaneseTokenizer UpperCamelCase = False UpperCamelCase = True def _lowercase ( self : Tuple ) -> Any: """simple docstring""" super().setUp() UpperCAmelCase = [ """[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは""", """世界""", """##世界""", """、""", """##、""", """。""", """##。""", ] UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _lowercase ( self : List[str] , __lowerCamelCase : int ) -> Dict: """simple docstring""" UpperCAmelCase = """こんにちは、世界。 \nこんばんは、世界。""" UpperCAmelCase = """こんにちは 、 世界 。 こんばんは 、 世界 。""" return input_text, output_text def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Dict ) -> int: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.get_input_output_texts(__lowerCamelCase ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.decode(__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase ) return text, ids def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" pass # TODO add if relevant def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" pass # TODO add if relevant def _lowercase ( self : Dict ) -> Dict: """simple docstring""" pass # TODO add if relevant def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file ) UpperCAmelCase = tokenizer.tokenize("""こんにちは、世界。\nこんばんは、世界。""" ) self.assertListEqual(__lowerCamelCase , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) def _lowercase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""mecab""" ) self.assertIsNotNone(__lowerCamelCase ) UpperCAmelCase = """こんにちは、世界。\nこんばんは、世界。""" UpperCAmelCase = tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) UpperCAmelCase = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(__lowerCamelCase , """wb""" ) as handle: pickle.dump(__lowerCamelCase , __lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as handle: UpperCAmelCase = pickle.load(__lowerCamelCase ) UpperCAmelCase = tokenizer_new.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = MecabTokenizer(mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def _lowercase ( self : Any ) -> Any: """simple docstring""" try: UpperCAmelCase = MecabTokenizer(mecab_dic="""unidic_lite""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" try: UpperCAmelCase = MecabTokenizer(mecab_dic="""unidic""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def _lowercase ( self : str ) -> Any: """simple docstring""" UpperCAmelCase = MecabTokenizer(do_lower_case=__lowerCamelCase , mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iphone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def _lowercase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" try: UpperCAmelCase = MecabTokenizer( do_lower_case=__lowerCamelCase , normalize_text=__lowerCamelCase , mecab_option="""-d /usr/local/lib/mecab/dic/jumandic""" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = MecabTokenizer(normalize_text=__lowerCamelCase , mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """ """, """。"""] , ) @require_sudachi def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""sudachi""" ) self.assertIsNotNone(__lowerCamelCase ) UpperCAmelCase = """こんにちは、世界。\nこんばんは、世界。""" UpperCAmelCase = tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) UpperCAmelCase = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(__lowerCamelCase , """wb""" ) as handle: pickle.dump(__lowerCamelCase , __lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as handle: UpperCAmelCase = pickle.load(__lowerCamelCase ) UpperCAmelCase = tokenizer_new.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) @require_sudachi def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" UpperCAmelCase = SudachiTokenizer(sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] , ) @require_sudachi def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""A""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国""", """人""", """参政""", """権"""] ) @require_sudachi def _lowercase ( self : List[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""B""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国人""", """参政権"""] ) @require_sudachi def _lowercase ( self : str ) -> List[str]: """simple docstring""" UpperCAmelCase = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""C""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国人参政権"""] ) @require_sudachi def _lowercase ( self : str ) -> Optional[int]: """simple docstring""" UpperCAmelCase = SudachiTokenizer(do_lower_case=__lowerCamelCase , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iphone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] , ) @require_sudachi def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = SudachiTokenizer(normalize_text=__lowerCamelCase , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """\u3000""", """。""", """ """, """ """] , ) @require_sudachi def _lowercase ( self : str ) -> Any: """simple docstring""" UpperCAmelCase = SudachiTokenizer(trim_whitespace=__lowerCamelCase , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) @require_jumanpp def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""jumanpp""" ) self.assertIsNotNone(__lowerCamelCase ) UpperCAmelCase = """こんにちは、世界。\nこんばんは、世界。""" UpperCAmelCase = tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) UpperCAmelCase = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(__lowerCamelCase , """wb""" ) as handle: pickle.dump(__lowerCamelCase , __lowerCamelCase ) with open(__lowerCamelCase , """rb""" ) as handle: UpperCAmelCase = pickle.load(__lowerCamelCase ) UpperCAmelCase = tokenizer_new.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) @require_jumanpp def _lowercase ( self : int ) -> Tuple: """simple docstring""" UpperCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def _lowercase ( self : List[Any] ) -> str: """simple docstring""" UpperCAmelCase = JumanppTokenizer(do_lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iphone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def _lowercase ( self : List[str] ) -> str: """simple docstring""" UpperCAmelCase = JumanppTokenizer(normalize_text=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""ア""", """ッ""", """フ""", """゚""", """ル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = JumanppTokenizer(trim_whitespace=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """。"""] , ) @require_jumanpp def _lowercase ( self : List[str] ) -> int: """simple docstring""" UpperCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("""ありがとうございますm(_ _)m見つけるのが大変です。""" ) , ["""ありがとう""", """ございます""", """m(_ _)m""", """見つける""", """の""", """が""", """大変です""", """。"""] , ) def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは"""] UpperCAmelCase = {} for i, token in enumerate(__lowerCamelCase ): UpperCAmelCase = i UpperCAmelCase = WordpieceTokenizer(vocab=__lowerCamelCase , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) , ["""こんにちは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは""" ) , ["""こん""", """##ばんは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは こんばんにちは こんにちは""" ) , ["""こん""", """##ばんは""", """[UNK]""", """こんにちは"""] ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = BertJapaneseTokenizer.from_pretrained("""nlp-waseda/roberta-base-japanese-with-auto-jumanpp""" ) UpperCAmelCase = tokenizer.subword_tokenizer UpperCAmelCase = subword_tokenizer.tokenize("""国境 の 長い トンネル を 抜ける と 雪国 であった 。""" ) self.assertListEqual(__lowerCamelCase , ["""▁国境""", """▁の""", """▁長い""", """▁トンネル""", """▁を""", """▁抜ける""", """▁と""", """▁雪""", """国""", """▁であった""", """▁。"""] ) UpperCAmelCase = subword_tokenizer.tokenize("""こんばんは こんばん にち は こんにちは""" ) self.assertListEqual(__lowerCamelCase , ["""▁こん""", """ばん""", """は""", """▁こん""", """ばん""", """▁に""", """ち""", """▁は""", """▁こんにちは"""] ) def _lowercase ( self : List[str] ) -> Any: """simple docstring""" UpperCAmelCase = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese""" ) UpperCAmelCase = tokenizer.encode("""ありがとう。""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode("""どういたしまして。""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase , __lowerCamelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = BertJapaneseTokenizer UpperCamelCase = False def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" super().setUp() UpperCAmelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _lowercase ( self : Dict , **__lowerCamelCase : Dict ) -> List[Any]: """simple docstring""" return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type="""character""" , **__lowerCamelCase ) def _lowercase ( self : List[str] , __lowerCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """こんにちは、世界。 \nこんばんは、世界。""" UpperCAmelCase = """こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。""" return input_text, output_text def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" pass # TODO add if relevant def _lowercase ( self : str ) -> Tuple: """simple docstring""" pass # TODO add if relevant def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" pass # TODO add if relevant def _lowercase ( self : int ) -> int: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type="""character""" ) UpperCAmelCase = tokenizer.tokenize("""こんにちは、世界。 \nこんばんは、世界。""" ) self.assertListEqual( __lowerCamelCase , ["""こ""", """ん""", """に""", """ち""", """は""", """、""", """世""", """界""", """。""", """こ""", """ん""", """ば""", """ん""", """は""", """、""", """世""", """界""", """。"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [3, 4, 5, 6, 7, 1_1, 9, 1_0, 1_2, 3, 4, 8, 4, 7, 1_1, 9, 1_0, 1_2] ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] UpperCAmelCase = {} for i, token in enumerate(__lowerCamelCase ): UpperCAmelCase = i UpperCAmelCase = CharacterTokenizer(vocab=__lowerCamelCase , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) , ["""こ""", """ん""", """に""", """ち""", """は"""] ) self.assertListEqual(tokenizer.tokenize("""こんにちほ""" ) , ["""こ""", """ん""", """に""", """ち""", """[UNK]"""] ) def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese-char""" ) UpperCAmelCase = tokenizer.encode("""ありがとう。""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode("""どういたしまして。""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase , __lowerCamelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowercase ( unittest.TestCase ): def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """cl-tohoku/bert-base-japanese""" UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) class __lowercase ( unittest.TestCase ): def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = """cl-tohoku/bert-base-japanese""" with self.assertLogs("""transformers""" , level="""WARNING""" ) as cm: BertTokenizer.from_pretrained(__lowerCamelCase ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) ) UpperCAmelCase = """bert-base-cased""" with self.assertLogs("""transformers""" , level="""WARNING""" ) as cm: BertJapaneseTokenizer.from_pretrained(__lowerCamelCase ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) )
711
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
712
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC __a = parse(importlib.metadata.version("""torch""")) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->str: if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) UpperCAmelCase = STR_OPERATION_TO_FUNC[operation] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = parse(importlib.metadata.version(lowerCAmelCase_ ) ) return operation(lowerCAmelCase_ , parse(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Any: return compare_versions(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )
713
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def _UpperCamelCase ( lowerCAmelCase_=None ) ->List[str]: UpperCAmelCase = argparse.ArgumentParser(add_help=lowerCAmelCase_ , allow_abbrev=lowerCAmelCase_ ) # The main config parser UpperCAmelCase = config_command_parser(lowerCAmelCase_ ) # The subparser to add commands to UpperCAmelCase = config_parser.add_subparsers(title="""subcommands""" , dest="""subcommand""" ) # Then add other parsers with the parent parser default_command_parser(lowerCAmelCase_ , parents=[parent_parser] ) update_command_parser(lowerCAmelCase_ , parents=[parent_parser] ) return config_parser def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = get_config_parser() UpperCAmelCase = config_parser.parse_args() if not hasattr(lowerCAmelCase_ , """func""" ): config_parser.print_help() exit(1 ) # Run args.func(lowerCAmelCase_ ) if __name__ == "__main__": main()
714
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
0
# Lint as: python3 import itertools import os import re __a = re.compile(R"""([A-Z]+)([A-Z][a-z])""") __a = re.compile(R"""([a-z\d])([A-Z])""") __a = re.compile(R"""(?<!_)_(?!_)""") __a = re.compile(R"""(_{2,})""") __a = R"""^\w+(\.\w+)*$""" __a = R"""<>:/\|?*""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Union[str, Any]: UpperCAmelCase = _uppercase_uppercase_re.sub(R"""\1_\2""" , lowerCAmelCase_ ) UpperCAmelCase = _lowercase_uppercase_re.sub(R"""\1_\2""" , lowerCAmelCase_ ) return name.lower() def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = _single_underscore_re.split(lowerCAmelCase_ ) UpperCAmelCase = [_multiple_underscores_re.split(lowerCAmelCase_ ) for n in name] return "".join(n.capitalize() for n in itertools.chain.from_iterable(lowerCAmelCase_ ) if n != """""" ) def _UpperCamelCase ( lowerCAmelCase_ ) ->Any: if os.path.basename(lowerCAmelCase_ ) != name: raise ValueError(F"""Should be a dataset name, not a path: {name}""" ) return camelcase_to_snakecase(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: if os.path.basename(lowerCAmelCase_ ) != name: raise ValueError(F"""Should be a dataset name, not a path: {name}""" ) if not re.match(_split_re , lowerCAmelCase_ ): raise ValueError(F"""Split name should match '{_split_re}'' but got '{split}'.""" ) return F"""{filename_prefix_for_name(lowerCAmelCase_ )}-{split}""" def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ) ->Optional[int]: UpperCAmelCase = filename_prefix_for_split(lowerCAmelCase_ , lowerCAmelCase_ ) if filetype_suffix: prefix += F""".{filetype_suffix}""" UpperCAmelCase = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) return F"""{filepath}*""" def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None ) ->Optional[Any]: UpperCAmelCase = filename_prefix_for_split(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) if shard_lengths: UpperCAmelCase = len(lowerCAmelCase_ ) UpperCAmelCase = [F"""{prefix}-{shard_id:05d}-of-{num_shards:05d}""" for shard_id in range(lowerCAmelCase_ )] if filetype_suffix: UpperCAmelCase = [filename + F""".{filetype_suffix}""" for filename in filenames] return filenames else: UpperCAmelCase = prefix if filetype_suffix: filename += F""".{filetype_suffix}""" return [filename]
715
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
0
import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow __a = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ """text-classification""", """language-modeling""", """summarization""", """token-classification""", """question-answering""", ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) __a = logging.getLogger() def _UpperCamelCase ( ) ->Union[str, Any]: UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""-f""" ) UpperCAmelCase = parser.parse_args() return args.f def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_="eval" ) ->Optional[int]: UpperCAmelCase = os.path.join(lowerCAmelCase_ , F"""{split}_results.json""" ) if os.path.exists(lowerCAmelCase_ ): with open(lowerCAmelCase_ , """r""" ) as f: return json.load(lowerCAmelCase_ ) raise ValueError(F"""can't find {path}""" ) __a = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class A ( __snake_case ): def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_flax_glue.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_accuracy"""] , 0.75 ) @slow def _lowercase ( self : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_clm_flax.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertLess(result["""eval_perplexity"""] , 1_0_0 ) @slow def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_summarization_flax.main() UpperCAmelCase = get_results(__lowerCamelCase , split="""test""" ) self.assertGreaterEqual(result["""test_rouge1"""] , 1_0 ) self.assertGreaterEqual(result["""test_rouge2"""] , 2 ) self.assertGreaterEqual(result["""test_rougeL"""] , 7 ) self.assertGreaterEqual(result["""test_rougeLsum"""] , 7 ) @slow def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_mlm_flax.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertLess(result["""eval_perplexity"""] , 4_2 ) @slow def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_ta_mlm_flax.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_accuracy"""] , 0.42 ) @slow def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = 7 if get_gpu_count() > 1 else 2 UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_flax_ner.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_accuracy"""] , 0.75 ) self.assertGreaterEqual(result["""eval_f1"""] , 0.3 ) @slow def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_qa.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_f1"""] , 3_0 ) self.assertGreaterEqual(result["""eval_exact"""] , 3_0 )
716
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
0
import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class __lowercase ( __snake_case ): UpperCamelCase = (EulerDiscreteScheduler,) UpperCamelCase = 10 def _lowercase ( self : Tuple , **__lowerCamelCase : int ) -> List[str]: """simple docstring""" UpperCAmelCase = { """num_train_timesteps""": 1_1_0_0, """beta_start""": 0.0_001, """beta_end""": 0.02, """beta_schedule""": """linear""", } config.update(**__lowerCamelCase ) return config def _lowercase ( self : List[Any] ) -> int: """simple docstring""" for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=__lowerCamelCase ) def _lowercase ( self : str ) -> str: """simple docstring""" for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=__lowerCamelCase , beta_end=__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=__lowerCamelCase ) def _lowercase ( self : int ) -> Tuple: """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma UpperCAmelCase = sample.to(__lowerCamelCase ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase = scheduler.scale_model_input(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase ) UpperCAmelCase = output.prev_sample UpperCAmelCase = torch.sum(torch.abs(__lowerCamelCase ) ) UpperCAmelCase = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 10.0_807 ) < 1e-2 assert abs(result_mean.item() - 0.0_131 ) < 1e-3 def _lowercase ( self : str ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config(prediction_type="""v_prediction""" ) UpperCAmelCase = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma UpperCAmelCase = sample.to(__lowerCamelCase ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase = scheduler.scale_model_input(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase ) UpperCAmelCase = output.prev_sample UpperCAmelCase = torch.sum(torch.abs(__lowerCamelCase ) ) UpperCAmelCase = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 0.0_002 ) < 1e-2 assert abs(result_mean.item() - 2.2_6_7_6e-0_6 ) < 1e-3 def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps , device=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() UpperCAmelCase = sample.to(__lowerCamelCase ) for t in scheduler.timesteps: UpperCAmelCase = scheduler.scale_model_input(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase ) UpperCAmelCase = output.prev_sample UpperCAmelCase = torch.sum(torch.abs(__lowerCamelCase ) ) UpperCAmelCase = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 10.0_807 ) < 1e-2 assert abs(result_mean.item() - 0.0_131 ) < 1e-3 def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**__lowerCamelCase , use_karras_sigmas=__lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps , device=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() UpperCAmelCase = sample.to(__lowerCamelCase ) for t in scheduler.timesteps: UpperCAmelCase = scheduler.scale_model_input(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase ) UpperCAmelCase = output.prev_sample UpperCAmelCase = torch.sum(torch.abs(__lowerCamelCase ) ) UpperCAmelCase = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 124.52_299_499_511_719 ) < 1e-2 assert abs(result_mean.item() - 0.16_213_932_633_399_963 ) < 1e-3
717
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
0
import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __a = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. __a = importlib.util.spec_from_file_location( """transformers""", os.path.join(PATH_TO_TRANSFORMERS, """__init__.py"""), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) __a = spec.loader.load_module() __a = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __a = re.compile("""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") __a = { """CLIPConfigMixin""", """DecisionTransformerConfigMixin""", """EncoderDecoderConfigMixin""", """RagConfigMixin""", """SpeechEncoderDecoderConfigMixin""", """VisionEncoderDecoderConfigMixin""", """VisionTextDualEncoderConfigMixin""", } def _UpperCamelCase ( ) ->int: UpperCAmelCase = [] for config_class in list(CONFIG_MAPPING.values() ): UpperCAmelCase = False # source code of `config_class` UpperCAmelCase = inspect.getsource(lowerCAmelCase_ ) UpperCAmelCase = _re_checkpoint.findall(lowerCAmelCase_ ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` UpperCAmelCase , UpperCAmelCase = checkpoint # verify the checkpoint name corresponds to the checkpoint link UpperCAmelCase = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: UpperCAmelCase = True break UpperCAmelCase = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: UpperCAmelCase = """\n""".join(sorted(lowerCAmelCase_ ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
718
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
0
'''simple docstring''' import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file UpperCAmelCase = TapasConfig.from_json_file(lowerCAmelCase_ ) # set absolute/relative position embeddings parameter UpperCAmelCase = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": UpperCAmelCase = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "WTQ": # run_task_main.py hparams UpperCAmelCase = 4 UpperCAmelCase = True # hparam_utils.py hparams UpperCAmelCase = 0.66_4694 UpperCAmelCase = 0.20_7951 UpperCAmelCase = 0.12_1194 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = 0.035_2513 UpperCAmelCase = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams UpperCAmelCase = 4 UpperCAmelCase = False # hparam_utils.py hparams UpperCAmelCase = 36.4519 UpperCAmelCase = 0.90_3421 UpperCAmelCase = 222.088 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 0.76_3141 UpperCAmelCase = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "TABFACT": UpperCAmelCase = TapasForSequenceClassification(config=lowerCAmelCase_ ) elif task == "MLM": UpperCAmelCase = TapasForMaskedLM(config=lowerCAmelCase_ ) elif task == "INTERMEDIATE_PRETRAINING": UpperCAmelCase = TapasModel(config=lowerCAmelCase_ ) else: raise ValueError(F"""Task {task} not supported.""" ) print(F"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Save pytorch-model (weights and configuration) print(F"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowerCAmelCase_ ) # Save tokenizer files print(F"""Save tokenizer files to {pytorch_dump_path}""" ) UpperCAmelCase = TapasTokenizer(vocab_file=tf_checkpoint_path[:-1_0] + """vocab.txt""" , model_max_length=5_1_2 ) tokenizer.save_pretrained(lowerCAmelCase_ ) print("""Used relative position embeddings:""" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __a = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
719
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
0
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = DDIMPipeline UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS UpperCamelCase = PipelineTesterMixin.required_optional_params - { '''num_images_per_prompt''', '''latents''', '''callback''', '''callback_steps''', } UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase = False def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) UpperCAmelCase = DDIMScheduler() UpperCAmelCase = {"""unet""": unet, """scheduler""": scheduler} return components def _lowercase ( self : int , __lowerCamelCase : Dict , __lowerCamelCase : Optional[Any]=0 ) -> Optional[Any]: """simple docstring""" if str(__lowerCamelCase ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(__lowerCamelCase ) else: UpperCAmelCase = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) UpperCAmelCase = { """batch_size""": 1, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = """cpu""" UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**__lowerCamelCase ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = self.get_dummy_inputs(__lowerCamelCase ) UpperCAmelCase = pipe(**__lowerCamelCase ).images UpperCAmelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 3_2, 3_2, 3) ) UpperCAmelCase = np.array( [1.0_0_0e0_0, 5.7_1_7e-0_1, 4.7_1_7e-0_1, 1.0_0_0e0_0, 0.0_0_0e0_0, 1.0_0_0e0_0, 3.0_0_0e-0_4, 0.0_0_0e0_0, 9.0_0_0e-0_4] ) UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__lowerCamelCase , 1e-3 ) def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" super().test_save_load_local(expected_max_difference=3e-3 ) def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3e-3 ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = """google/ddpm-cifar10-32""" UpperCAmelCase = UNetaDModel.from_pretrained(__lowerCamelCase ) UpperCAmelCase = DDIMScheduler() UpperCAmelCase = DDIMPipeline(unet=__lowerCamelCase , scheduler=__lowerCamelCase ) ddim.to(__lowerCamelCase ) ddim.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = ddim(generator=__lowerCamelCase , eta=0.0 , output_type="""numpy""" ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) UpperCAmelCase = np.array([0.1_723, 0.1_617, 0.1_600, 0.1_626, 0.1_497, 0.1_513, 0.1_505, 0.1_442, 0.1_453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" UpperCAmelCase = """google/ddpm-ema-bedroom-256""" UpperCAmelCase = UNetaDModel.from_pretrained(__lowerCamelCase ) UpperCAmelCase = DDIMScheduler.from_pretrained(__lowerCamelCase ) UpperCAmelCase = DDIMPipeline(unet=__lowerCamelCase , scheduler=__lowerCamelCase ) ddpm.to(__lowerCamelCase ) ddpm.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = ddpm(generator=__lowerCamelCase , output_type="""numpy""" ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 2_5_6, 2_5_6, 3) UpperCAmelCase = np.array([0.0_060, 0.0_201, 0.0_344, 0.0_024, 0.0_018, 0.0_002, 0.0_022, 0.0_000, 0.0_069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
720
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """uw-madison/mra-base-512-4""": """https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''mra''' def __init__( self : List[Any] , __lowerCamelCase : Dict=5_0_2_6_5 , __lowerCamelCase : Tuple=7_6_8 , __lowerCamelCase : int=1_2 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : str=3_0_7_2 , __lowerCamelCase : str="gelu" , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : int=5_1_2 , __lowerCamelCase : List[str]=1 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=1e-5 , __lowerCamelCase : List[str]="absolute" , __lowerCamelCase : str=4 , __lowerCamelCase : List[str]="full" , __lowerCamelCase : Tuple=0 , __lowerCamelCase : Any=0 , __lowerCamelCase : Dict=1 , __lowerCamelCase : Optional[int]=0 , __lowerCamelCase : Union[str, Any]=2 , **__lowerCamelCase : List[Any] , ) -> List[Any]: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = type_vocab_size UpperCAmelCase = layer_norm_eps UpperCAmelCase = position_embedding_type UpperCAmelCase = block_per_row UpperCAmelCase = approx_mode UpperCAmelCase = initial_prior_first_n_blocks UpperCAmelCase = initial_prior_diagonal_n_blocks
721
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
0
from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowercase : def __init__( self : Any , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any=3 , __lowerCamelCase : int=3_2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Dict=1_0 , __lowerCamelCase : List[Any]=[1_0, 2_0, 3_0, 4_0] , __lowerCamelCase : int=[1, 1, 2, 1] , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : int="relu" , __lowerCamelCase : Dict=3 , __lowerCamelCase : List[str]=None , ) -> int: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = num_channels UpperCAmelCase = embeddings_size UpperCAmelCase = hidden_sizes UpperCAmelCase = depths UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_act UpperCAmelCase = num_labels UpperCAmelCase = scope UpperCAmelCase = len(__lowerCamelCase ) def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Any ) -> Any: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def _lowercase ( self : Dict , __lowerCamelCase : Dict , __lowerCamelCase : Dict , __lowerCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFRegNetModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , training=__lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def _lowercase ( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Any ) -> List[str]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = TFRegNetForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase , training=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : List[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFRegNetModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def _lowercase ( self : str ) -> List[Any]: """simple docstring""" super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" pass def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" def check_hidden_states_output(__lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] ): UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) , training=__lowerCamelCase ) UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(__lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCAmelCase = layer_type UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Dict: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(__lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : List[Any] , __lowerCamelCase : List[str]={} ): UpperCAmelCase = model(__lowerCamelCase , return_dict=__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , return_dict=__lowerCamelCase , **__lowerCamelCase ).to_tuple() def recursive_check(__lowerCamelCase : List[Any] , __lowerCamelCase : Dict ): if isinstance(__lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(__lowerCamelCase , __lowerCamelCase ): recursive_check(__lowerCamelCase , __lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(__lowerCamelCase , __lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F""" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}""" ) , ) recursive_check(__lowerCamelCase , __lowerCamelCase ) for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , {"""output_hidden_states""": True} ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , {"""output_hidden_states""": True} ) def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _lowercase ( self : List[Any] ) -> str: """simple docstring""" for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFRegNetModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Union[str, Any]: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : str ) -> Tuple: """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" UpperCAmelCase = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase , training=__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-0.4_180, -1.5_051, -3.4_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 )
700
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
0
def _UpperCamelCase ( lowerCAmelCase_ ) ->list[int]: if length <= 0 or not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError("""Length must be a positive integer.""" ) return [n * (2 * n - 1) for n in range(lowerCAmelCase_ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
701
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
0
from math import isqrt, loga def _UpperCamelCase ( lowerCAmelCase_ ) ->list[int]: UpperCAmelCase = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = False return [i for i in range(2 , lowerCAmelCase_ ) if is_prime[i]] def _UpperCamelCase ( lowerCAmelCase_ = 8_0_0_8_0_0 , lowerCAmelCase_ = 8_0_0_8_0_0 ) ->int: UpperCAmelCase = degree * loga(lowerCAmelCase_ ) UpperCAmelCase = int(lowerCAmelCase_ ) UpperCAmelCase = calculate_prime_numbers(lowerCAmelCase_ ) UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = len(lowerCAmelCase_ ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(F"""{solution() = }""")
702
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
0
from itertools import count def _UpperCamelCase ( lowerCAmelCase_ = 5_0 ) ->int: UpperCAmelCase = [1] * min_block_length for n in count(lowerCAmelCase_ ): fill_count_functions.append(1 ) for block_length in range(lowerCAmelCase_ , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 1_0_0_0_0_0_0: break return n if __name__ == "__main__": print(F"""{solution() = }""")
703
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
0
import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset __a = """bert-base-cased""" __a = """google/pegasus-xsum""" __a = [""" Sam ate lunch today.""", """Sams lunch ingredients."""] __a = ["""A very interesting story about what I ate for lunch.""", """Avocado, celery, turkey, coffee"""] __a = """patrickvonplaten/t5-tiny-random""" __a = """sshleifer/bart-tiny-random""" __a = """sshleifer/tiny-mbart""" __a = """sshleifer/tiny-marian-en-de""" def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: UpperCAmelCase = """\n""".join(lowerCAmelCase_ ) Path(lowerCAmelCase_ ).open("""w""" ).writelines(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->str: for split in ["train", "val", "test"]: _dump_articles(os.path.join(lowerCAmelCase_ , F"""{split}.source""" ) , lowerCAmelCase_ ) _dump_articles(os.path.join(lowerCAmelCase_ , F"""{split}.target""" ) , lowerCAmelCase_ ) return tmp_dir class __lowercase ( __snake_case ): @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) @slow def _lowercase ( self : Optional[int] , __lowerCamelCase : Any ) -> List[str]: """simple docstring""" UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) UpperCAmelCase = max(len(tokenizer.encode(__lowerCamelCase ) ) for a in ARTICLES ) UpperCAmelCase = max(len(tokenizer.encode(__lowerCamelCase ) ) for a in SUMMARIES ) UpperCAmelCase = 4 UpperCAmelCase = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated UpperCAmelCase , UpperCAmelCase = """ro_RO""", """de_DE""" # ignored for all but mbart, but never causes error. UpperCAmelCase = SeqaSeqDataset( __lowerCamelCase , data_dir=__lowerCamelCase , type_path="""train""" , max_source_length=__lowerCamelCase , max_target_length=__lowerCamelCase , src_lang=__lowerCamelCase , tgt_lang=__lowerCamelCase , ) UpperCAmelCase = DataLoader(__lowerCamelCase , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(__lowerCamelCase , __lowerCamelCase ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place UpperCAmelCase = shift_tokens_right(batch["""labels"""] , tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) UpperCAmelCase = max(len(tokenizer.encode(__lowerCamelCase ) ) for a in ARTICLES ) UpperCAmelCase = max(len(tokenizer.encode(__lowerCamelCase ) ) for a in SUMMARIES ) UpperCAmelCase = 4 UpperCAmelCase = LegacySeqaSeqDataset( __lowerCamelCase , data_dir=__lowerCamelCase , type_path="""train""" , max_source_length=2_0 , max_target_length=__lowerCamelCase , ) UpperCAmelCase = DataLoader(__lowerCamelCase , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 2_0 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = AutoTokenizer.from_pretrained("""facebook/mbart-large-cc25""" ) UpperCAmelCase = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) UpperCAmelCase = tmp_dir.joinpath("""train.source""" ).open().readlines() UpperCAmelCase = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(__lowerCamelCase , __lowerCamelCase , 1_2_8 , __lowerCamelCase ) UpperCAmelCase = {x.name for x in tmp_dir.iterdir()} UpperCAmelCase = {x.name for x in save_dir.iterdir()} UpperCAmelCase = save_dir.joinpath("""train.source""" ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(__lowerCamelCase ) < len(__lowerCamelCase ) assert len(__lowerCamelCase ) == 1 assert len(packed_examples[0] ) == sum(len(__lowerCamelCase ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE , reason="""This test requires fairseq""" ) def _lowercase ( self : Tuple ) -> str: """simple docstring""" if not FAIRSEQ_AVAILABLE: return UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self._get_dataset(max_len=6_4 ) UpperCAmelCase = 6_4 UpperCAmelCase = ds.make_dynamic_sampler(__lowerCamelCase , required_batch_size_multiple=__lowerCamelCase ) UpperCAmelCase = [len(__lowerCamelCase ) for x in batch_sampler] assert len(set(__lowerCamelCase ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(__lowerCamelCase ) == len(__lowerCamelCase ) # no dropped or added examples UpperCAmelCase = DataLoader(__lowerCamelCase , batch_sampler=__lowerCamelCase , collate_fn=ds.collate_fn , num_workers=2 ) UpperCAmelCase = [] UpperCAmelCase = [] for batch in data_loader: UpperCAmelCase = batch["""input_ids"""].shape UpperCAmelCase = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple UpperCAmelCase = np.product(batch["""input_ids"""].shape ) num_src_per_batch.append(__lowerCamelCase ) if num_src_tokens > (max_tokens * 1.1): failures.append(__lowerCamelCase ) assert num_src_per_batch[0] == max(__lowerCamelCase ) if failures: raise AssertionError(F"""too many tokens in {len(__lowerCamelCase )} batches""" ) def _lowercase ( self : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self._get_dataset(max_len=5_1_2 ) UpperCAmelCase = 2 UpperCAmelCase = ds.make_sortish_sampler(__lowerCamelCase , shuffle=__lowerCamelCase ) UpperCAmelCase = DataLoader(__lowerCamelCase , batch_size=__lowerCamelCase , collate_fn=ds.collate_fn , num_workers=2 ) UpperCAmelCase = DataLoader(__lowerCamelCase , batch_size=__lowerCamelCase , collate_fn=ds.collate_fn , num_workers=2 , sampler=__lowerCamelCase ) UpperCAmelCase = tokenizer.pad_token_id def count_pad_tokens(__lowerCamelCase : str , __lowerCamelCase : int="input_ids" ): return [batch[k].eq(__lowerCamelCase ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(__lowerCamelCase , k="""labels""" ) ) < sum(count_pad_tokens(__lowerCamelCase , k="""labels""" ) ) assert sum(count_pad_tokens(__lowerCamelCase ) ) < sum(count_pad_tokens(__lowerCamelCase ) ) assert len(__lowerCamelCase ) == len(__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : int=1_0_0_0 , __lowerCamelCase : Dict=1_2_8 ) -> Optional[Any]: """simple docstring""" if os.getenv("""USE_REAL_DATA""" , __lowerCamelCase ): UpperCAmelCase = """examples/seq2seq/wmt_en_ro""" UpperCAmelCase = max_len * 2 * 6_4 if not Path(__lowerCamelCase ).joinpath("""train.len""" ).exists(): save_len_file(__lowerCamelCase , __lowerCamelCase ) else: UpperCAmelCase = """examples/seq2seq/test_data/wmt_en_ro""" UpperCAmelCase = max_len * 4 save_len_file(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = SeqaSeqDataset( __lowerCamelCase , data_dir=__lowerCamelCase , type_path="""train""" , max_source_length=__lowerCamelCase , max_target_length=__lowerCamelCase , n_obs=__lowerCamelCase , ) return ds, max_tokens, tokenizer def _lowercase ( self : str ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self._get_dataset() UpperCAmelCase = set(DistributedSortishSampler(__lowerCamelCase , 2_5_6 , num_replicas=2 , rank=0 , add_extra_examples=__lowerCamelCase ) ) UpperCAmelCase = set(DistributedSortishSampler(__lowerCamelCase , 2_5_6 , num_replicas=2 , rank=1 , add_extra_examples=__lowerCamelCase ) ) assert idsa.intersection(__lowerCamelCase ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) def _lowercase ( self : Tuple , __lowerCamelCase : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase , use_fast=__lowerCamelCase ) if tok_name == MBART_TINY: UpperCAmelCase = SeqaSeqDataset( __lowerCamelCase , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path="""train""" , max_source_length=4 , max_target_length=8 , src_lang="""EN""" , tgt_lang="""FR""" , ) UpperCAmelCase = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: UpperCAmelCase = SeqaSeqDataset( __lowerCamelCase , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path="""train""" , max_source_length=4 , max_target_length=8 , ) UpperCAmelCase = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(__lowerCamelCase ) == 1 if tok_name == BART_TINY else len(__lowerCamelCase ) == 0
704
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
0
'''simple docstring''' import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = FlaxAutoencoderKL @property def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" UpperCAmelCase = 4 UpperCAmelCase = 3 UpperCAmelCase = (3_2, 3_2) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = jax.random.uniform(__lowerCamelCase , ((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = { """block_out_channels""": [3_2, 6_4], """in_channels""": 3, """out_channels""": 3, """down_block_types""": ["""DownEncoderBlock2D""", """DownEncoderBlock2D"""], """up_block_types""": ["""UpDecoderBlock2D""", """UpDecoderBlock2D"""], """latent_channels""": 4, } UpperCAmelCase = self.dummy_input return init_dict, inputs_dict
705
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
0
'''simple docstring''' import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
706
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''lxmert''' UpperCamelCase = {} def __init__( self : Optional[int] , __lowerCamelCase : Dict=3_0_5_2_2 , __lowerCamelCase : List[Any]=7_6_8 , __lowerCamelCase : Any=1_2 , __lowerCamelCase : Optional[Any]=9_5_0_0 , __lowerCamelCase : Union[str, Any]=1_6_0_0 , __lowerCamelCase : Union[str, Any]=4_0_0 , __lowerCamelCase : List[Any]=3_0_7_2 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : List[str]=5_1_2 , __lowerCamelCase : int=2 , __lowerCamelCase : Any=0.02 , __lowerCamelCase : Union[str, Any]=1e-1_2 , __lowerCamelCase : List[Any]=9 , __lowerCamelCase : Tuple=5 , __lowerCamelCase : List[str]=5 , __lowerCamelCase : Optional[Any]=2_0_4_8 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : List[str]=6.67 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : int=True , **__lowerCamelCase : Dict , ) -> Any: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = num_qa_labels UpperCAmelCase = num_object_labels UpperCAmelCase = num_attr_labels UpperCAmelCase = l_layers UpperCAmelCase = x_layers UpperCAmelCase = r_layers UpperCAmelCase = visual_feat_dim UpperCAmelCase = visual_pos_dim UpperCAmelCase = visual_loss_normalizer UpperCAmelCase = task_matched UpperCAmelCase = task_mask_lm UpperCAmelCase = task_obj_predict UpperCAmelCase = task_qa UpperCAmelCase = visual_obj_loss UpperCAmelCase = visual_attr_loss UpperCAmelCase = visual_feat_loss UpperCAmelCase = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**__lowerCamelCase )
707
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
0
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __a = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = XLMProphetNetTokenizer UpperCamelCase = False UpperCamelCase = True def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMProphetNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = """[PAD]""" UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """[PAD]""" ) self.assertEqual(vocab_keys[1] , """[CLS]""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(__lowerCamelCase ) , 1_0_1_2 ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_2 ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = XLMProphetNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__lowerCamelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, -9, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, -9, 4] ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """[UNK]""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """[UNK]""", """.""", ] , ) @cached_property def _lowercase ( self : int ) -> Any: """simple docstring""" return XLMProphetNetTokenizer.from_pretrained("""microsoft/xprophetnet-large-wiki100-cased""" ) @slow def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [3_5_3_8_9, 6_6_7_2, 4_9, 2] self.assertListEqual(__lowerCamelCase , self.big_tokenizer.encode(__lowerCamelCase ) ) @slow def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[1_1_0_7_3, 8_2_7_8_3, 1_8, 2_6, 8_2_7_8_3, 5_4_9, 5_1_5_4_0, 2_4_8, 1_7_2_0_9, 1_3_0_1, 2_1_7, 2_0, 2_1_5_1_8_6, 1_3_2_5, 1_4_7, 1_7_2_0_9, 1_3_0_1, 2_1_7, 2_0, 5_6_3_7_0, 5_3, 1_2_2_0_2_0, 2_0, 1_6_4_7_7, 2_7, 8_7_3_5_5, 4_5_4_8, 2_0, 4_7_2_8, 7_8_3_9_2, 1_7, 1_5_9_9_6_9, 1_8, 2_6, 2_4_4_9_1, 6_2_9, 1_5, 5_3_8, 2_2_7_0_4, 5_4_3_9, 1_5, 2_7_8_8, 2_4_4_9_1, 9_8_8_5, 1_5, 4_3_5_3_4, 6_0_5, 1_5, 8_1_4, 1_8_4_0_3, 3_3_2_0_0, 2_9, 1_5, 4_3_5_3_4, 2_4_4_5_8, 1_2_4_1_0, 1_1_1, 2_4_9_6_6, 8_3_6_6_9, 9_6_3_7, 1_4_4_0_6_8, 2_6, 8_5_0, 2_2_3_4_6, 2_7, 1_4_7, 2_4_9_6_6, 8_3_6_6_9, 8_3_4_9_0, 2_6, 3_9_1_1_3, 7_3_5, 2_7, 6_8_9, 6_5_6, 2_8_0_0, 1_3_3_9, 4_6_0_0, 5_3, 1_2_2_0_2_0, 1_1_5_7_8_5, 3_4, 8_1_6, 1_3_3_9, 4_6_8_8_7, 1_8, 1_4_7, 5_3_9_0_5, 1_9_5_1, 4_2_2_3_8, 4_1_1_7_0, 1_7_7_3_2, 8_3_4, 4_3_6, 1_5, 2_7_5_2_3, 9_8_7_3_3, 2_1_7, 1_4_7, 5_5_4_2, 4_9_8_1, 9_3_0, 1_7_3_4_7, 1_6, 2], [2_0_0_9_1, 6_2_9, 9_4, 8_2_7_8_6, 5_8, 4_9_0, 2_0, 1_5_2_8, 8_4, 5_3_9_0_5, 3_4_4, 8_0_5_9_2, 1_1_0_1_2_8, 1_8_8_2_2, 5_2_6_7, 1_3_0_6, 6_2, 1_5_2_5_3_7, 3_0_8, 7_9_9_7, 4_0_1, 1_2_4_4_2_7, 5_4_9, 3_5_4_4_2, 2_2_5, 1_0_9, 1_5_0_5_5, 2_5_7_4_8, 1_4_7, 7_1_1_9, 4_3_7_1_2, 3_4, 7_6_7, 1_3_5_3_6_6, 1_8, 1_6, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_9_2, 6_3_7_8_4, 1_1_9_4_6_6, 1_7, 1_4_7_8_0_8, 8_8_2_1_4, 1_8, 6_5_6, 8_1, 3_2, 3_2_9_6, 1_0_2_8_0, 1_6, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name="""microsoft/xprophetnet-large-wiki100-cased""" , revision="""1acad1643ddd54a44df6a1b797ada8373685d90e""" , )
708
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
0
from argparse import ArgumentParser from . import BaseTransformersCLICommand def _UpperCamelCase ( lowerCAmelCase_ ): return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class __lowercase ( __snake_case ): @staticmethod def _lowercase ( __lowerCamelCase : ArgumentParser ) -> Tuple: """simple docstring""" UpperCAmelCase = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" , type=__lowerCamelCase , default=__lowerCamelCase , help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" , action="""store_true""" , help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" , action="""store_true""" , help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" , ) download_parser.add_argument("""model""" , type=__lowerCamelCase , help="""Name of the model to download""" ) download_parser.set_defaults(func=__lowerCamelCase ) def __init__( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : bool , __lowerCamelCase : bool ) -> Any: """simple docstring""" UpperCAmelCase = model UpperCAmelCase = cache UpperCAmelCase = force UpperCAmelCase = trust_remote_code def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
709
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
0
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return number | (1 << position) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return number & ~(1 << position) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return number ^ (1 << position) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: return ((number >> position) & 1) == 1 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
710
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer __a = logging.get_logger(__name__) __a = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} __a = { """vocab_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json""" ), }, } __a = { """yjernite/retribert-base-uncased""": 512, } __a = { """yjernite/retribert-base-uncased""": {"""do_lower_case""": True}, } class __lowercase ( __snake_case ): UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = PRETRAINED_INIT_CONFIGURATION UpperCamelCase = RetriBertTokenizer UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self : Optional[int] , __lowerCamelCase : Tuple=None , __lowerCamelCase : str=None , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : int="[UNK]" , __lowerCamelCase : Dict="[SEP]" , __lowerCamelCase : Union[str, Any]="[PAD]" , __lowerCamelCase : Optional[Any]="[CLS]" , __lowerCamelCase : str="[MASK]" , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : int=None , **__lowerCamelCase : Any , ) -> Any: """simple docstring""" super().__init__( __lowerCamelCase , tokenizer_file=__lowerCamelCase , do_lower_case=__lowerCamelCase , unk_token=__lowerCamelCase , sep_token=__lowerCamelCase , pad_token=__lowerCamelCase , cls_token=__lowerCamelCase , mask_token=__lowerCamelCase , tokenize_chinese_chars=__lowerCamelCase , strip_accents=__lowerCamelCase , **__lowerCamelCase , ) UpperCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , __lowerCamelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , __lowerCamelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , __lowerCamelCase ) != tokenize_chinese_chars ): UpperCAmelCase = getattr(__lowerCamelCase , normalizer_state.pop("""type""" ) ) UpperCAmelCase = do_lower_case UpperCAmelCase = strip_accents UpperCAmelCase = tokenize_chinese_chars UpperCAmelCase = normalizer_class(**__lowerCamelCase ) UpperCAmelCase = do_lower_case def _lowercase ( self : int , __lowerCamelCase : str , __lowerCamelCase : Dict=None ) -> Optional[int]: """simple docstring""" UpperCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self : Optional[Any] , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase = self._tokenizer.model.save(__lowerCamelCase , name=__lowerCamelCase ) return tuple(__lowerCamelCase )
711
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = (1 - _cos) / 2 UpperCAmelCase = 1 - _cos UpperCAmelCase = 1 + alpha UpperCAmelCase = -2 * _cos UpperCAmelCase = 1 - alpha UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = (1 + _cos) / 2 UpperCAmelCase = -1 - _cos UpperCAmelCase = 1 + alpha UpperCAmelCase = -2 * _cos UpperCAmelCase = 1 - alpha UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = _sin / 2 UpperCAmelCase = 0 UpperCAmelCase = -ba UpperCAmelCase = 1 + alpha UpperCAmelCase = -2 * _cos UpperCAmelCase = 1 - alpha UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = 1 - alpha UpperCAmelCase = -2 * _cos UpperCAmelCase = 1 + alpha UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) , ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = 1_0 ** (gain_db / 4_0) UpperCAmelCase = 1 + alpha * big_a UpperCAmelCase = -2 * _cos UpperCAmelCase = 1 - alpha * big_a UpperCAmelCase = 1 + alpha / big_a UpperCAmelCase = -2 * _cos UpperCAmelCase = 1 - alpha / big_a UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) , ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = 1_0 ** (gain_db / 4_0) UpperCAmelCase = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase = 2 * sqrt(lowerCAmelCase_ ) * alpha UpperCAmelCase = big_a * (pmc + aaa) UpperCAmelCase = 2 * big_a * mpc UpperCAmelCase = big_a * (pmc - aaa) UpperCAmelCase = ppmc + aaa UpperCAmelCase = -2 * pmpc UpperCAmelCase = ppmc - aaa UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 / sqrt(2 ) , ) ->IIRFilter: UpperCAmelCase = tau * frequency / samplerate UpperCAmelCase = sin(lowerCAmelCase_ ) UpperCAmelCase = cos(lowerCAmelCase_ ) UpperCAmelCase = _sin / (2 * q_factor) UpperCAmelCase = 1_0 ** (gain_db / 4_0) UpperCAmelCase = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase = 2 * sqrt(lowerCAmelCase_ ) * alpha UpperCAmelCase = big_a * (ppmc + aaa) UpperCAmelCase = -2 * big_a * pmpc UpperCAmelCase = big_a * (ppmc - aaa) UpperCAmelCase = pmc + aaa UpperCAmelCase = 2 * mpc UpperCAmelCase = pmc - aaa UpperCAmelCase = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
712
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
from __future__ import annotations from typing import Any class __lowercase : def __init__( self : List[Any] , __lowerCamelCase : int ) -> None: """simple docstring""" UpperCAmelCase = num_of_nodes UpperCAmelCase = [] UpperCAmelCase = {} def _lowercase ( self : str , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int ) -> None: """simple docstring""" self.m_edges.append([u_node, v_node, weight] ) def _lowercase ( self : List[Any] , __lowerCamelCase : int ) -> int: """simple docstring""" if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int ) -> None: """simple docstring""" if self.m_component[u_node] != u_node: for k in self.m_component: UpperCAmelCase = self.find_component(__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : int ) -> None: """simple docstring""" if component_size[u_node] <= component_size[v_node]: UpperCAmelCase = v_node component_size[v_node] += component_size[u_node] self.set_component(__lowerCamelCase ) elif component_size[u_node] >= component_size[v_node]: UpperCAmelCase = self.find_component(__lowerCamelCase ) component_size[u_node] += component_size[v_node] self.set_component(__lowerCamelCase ) def _lowercase ( self : Any ) -> None: """simple docstring""" UpperCAmelCase = [] UpperCAmelCase = 0 UpperCAmelCase = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) UpperCAmelCase = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = edge UpperCAmelCase = self.m_component[u] UpperCAmelCase = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): UpperCAmelCase = [u, v, w] for edge in minimum_weight_edge: if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = edge UpperCAmelCase = self.m_component[u] UpperCAmelCase = self.m_component[v] if u_component != v_component: mst_weight += w self.union(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) print(F"""Added edge [{u} - {v}]\nAdded weight: {w}\n""" ) num_of_components -= 1 UpperCAmelCase = [-1] * self.m_num_of_nodes print(F"""The total weight of the minimal spanning tree is: {mst_weight}""" ) def _UpperCamelCase ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
713
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
0
def _UpperCamelCase ( lowerCAmelCase_ = 5_0 ) ->int: UpperCAmelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
714
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
0
from __future__ import annotations __a = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->tuple[list[list[int]], list[list[int]]]: UpperCAmelCase = [ [0 for col in range(len(grid[0] ) )] for row in range(len(lowerCAmelCase_ ) ) ] # the reference grid UpperCAmelCase = 1 UpperCAmelCase = [ [0 for col in range(len(grid[0] ) )] for row in range(len(lowerCAmelCase_ ) ) ] # the action grid UpperCAmelCase = init[0] UpperCAmelCase = init[1] UpperCAmelCase = 0 UpperCAmelCase = g + heuristic[x][y] # cost from starting cell to destination cell UpperCAmelCase = [[f, g, x, y]] UpperCAmelCase = False # flag that is set when search is complete UpperCAmelCase = False # flag set if we can't find expand while not found and not resign: if len(lowerCAmelCase_ ) == 0: raise ValueError("""Algorithm is unable to find solution""" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() UpperCAmelCase = cell.pop() UpperCAmelCase = next_cell[2] UpperCAmelCase = next_cell[3] UpperCAmelCase = next_cell[1] if x == goal[0] and y == goal[1]: UpperCAmelCase = True else: for i in range(len(lowerCAmelCase_ ) ): # to try out different valid actions UpperCAmelCase = x + DIRECTIONS[i][0] UpperCAmelCase = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(lowerCAmelCase_ ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: UpperCAmelCase = g + cost UpperCAmelCase = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) UpperCAmelCase = 1 UpperCAmelCase = i UpperCAmelCase = [] UpperCAmelCase = goal[0] UpperCAmelCase = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: UpperCAmelCase = x - DIRECTIONS[action[x][y]][0] UpperCAmelCase = y - DIRECTIONS[action[x][y]][1] UpperCAmelCase = xa UpperCAmelCase = ya invpath.append([x, y] ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): path.append(invpath[len(lowerCAmelCase_ ) - 1 - i] ) return path, action if __name__ == "__main__": __a = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] __a = [0, 0] # all coordinates are given in format [y,x] __a = [len(grid) - 1, len(grid[0]) - 1] __a = 1 # the cost map which pushes the path closer to the goal __a = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): __a = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map __a = 99 __a , __a = search(grid, init, goal, cost, heuristic) print("""ACTION MAP""") for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
715
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
0
from numpy import exp, pi, sqrt def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = 0.0 , lowerCAmelCase_ = 1.0 ) ->int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
716
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
0
def _UpperCamelCase ( lowerCAmelCase_ ) ->str: return "".join([hex(lowerCAmelCase_ )[2:].zfill(2 ).upper() for byte in list(lowerCAmelCase_ )] ) def _UpperCamelCase ( lowerCAmelCase_ ) ->bytes: # Check data validity, following RFC3548 # https://www.ietf.org/rfc/rfc3548.txt if (len(lowerCAmelCase_ ) % 2) != 0: raise ValueError( """Base16 encoded data is invalid: Data does not have an even number of hex digits.""" ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(lowerCAmelCase_ ) <= set("""0123456789ABCDEF""" ): raise ValueError( """Base16 encoded data is invalid: Data is not uppercase hex or it contains invalid characters.""" ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 1_6 ) for i in range(0 , len(lowerCAmelCase_ ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
717
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
0
# Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union __a = re.compile(R"""^(?P<major>\d+)""" R"""\.(?P<minor>\d+)""" R"""\.(?P<patch>\d+)$""") @total_ordering @dataclass class __lowercase : UpperCamelCase = 42 UpperCamelCase = None UpperCamelCase = None UpperCamelCase = None UpperCamelCase = None def _lowercase ( self : List[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = _str_to_version_tuple(self.version_str ) def __repr__( self : Optional[int] ) -> Dict: """simple docstring""" return F"""{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}""" @property def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" return self.major, self.minor, self.patch def _lowercase ( self : Dict , __lowerCamelCase : int ) -> Dict: """simple docstring""" if isinstance(__lowerCamelCase , __lowerCamelCase ): return Version(__lowerCamelCase ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): return other raise TypeError(F"""{other} (type {type(__lowerCamelCase )}) cannot be compared to version.""" ) def __eq__( self : int , __lowerCamelCase : List[Any] ) -> Optional[Any]: """simple docstring""" try: UpperCAmelCase = self._validate_operand(__lowerCamelCase ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self : Union[str, Any] , __lowerCamelCase : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self._validate_operand(__lowerCamelCase ) return self.tuple < other.tuple def __hash__( self : Optional[Any] ) -> Any: """simple docstring""" return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def _lowercase ( cls : List[str] , __lowerCamelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def _lowercase ( self : Optional[Any] ) -> str: """simple docstring""" return self.version_str def _UpperCamelCase ( lowerCAmelCase_ ) ->Any: UpperCAmelCase = _VERSION_REG.match(lowerCAmelCase_ ) if not res: raise ValueError(F"""Invalid version '{version_str}'. Format should be x.y.z with {{x,y,z}} being digits.""" ) return tuple(int(lowerCAmelCase_ ) for v in [res.group("""major""" ), res.group("""minor""" ), res.group("""patch""" )] ) def _UpperCamelCase ( lowerCAmelCase_ ) ->List[Any]: return ".".join(str(lowerCAmelCase_ ) for v in version_tuple )
718
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging __a = logging.get_logger(__name__) if is_vision_available(): import PIL class __lowercase ( __snake_case ): UpperCamelCase = ['''pixel_values'''] def __init__( self : List[str] , __lowerCamelCase : bool = True , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : PILImageResampling = PILImageResampling.BICUBIC , __lowerCamelCase : bool = True , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : bool = True , __lowerCamelCase : Union[int, float] = 1 / 2_5_5 , __lowerCamelCase : bool = True , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : bool = True , **__lowerCamelCase : Optional[Any] , ) -> None: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_2_4} UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase ) UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 2_2_4, """width""": 2_2_4} UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = resample UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCAmelCase = image_std if image_std is not None else OPENAI_CLIP_STD UpperCAmelCase = do_convert_rgb def _lowercase ( self : Dict , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : PILImageResampling = PILImageResampling.BICUBIC , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Tuple , ) -> np.ndarray: """simple docstring""" UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase = get_resize_output_image_size(__lowerCamelCase , size=size["""shortest_edge"""] , default_to_square=__lowerCamelCase ) return resize(__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Any , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Optional[int] , ) -> np.ndarray: """simple docstring""" UpperCAmelCase = get_size_dict(__lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(__lowerCamelCase , size=(size["""height"""], size["""width"""]) , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[int, float] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Dict , ) -> List[Any]: """simple docstring""" return rescale(__lowerCamelCase , scale=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[float, List[float]] , __lowerCamelCase : Union[float, List[float]] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Dict , ) -> np.ndarray: """simple docstring""" return normalize(__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : ImageInput , __lowerCamelCase : bool = None , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : PILImageResampling = None , __lowerCamelCase : bool = None , __lowerCamelCase : int = None , __lowerCamelCase : bool = None , __lowerCamelCase : float = None , __lowerCamelCase : bool = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : bool = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , __lowerCamelCase : Optional[ChannelDimension] = ChannelDimension.FIRST , **__lowerCamelCase : Dict , ) -> PIL.Image.Image: """simple docstring""" UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(__lowerCamelCase , param_name="""size""" , default_to_square=__lowerCamelCase ) UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase = crop_size if crop_size is not None else self.crop_size UpperCAmelCase = get_size_dict(__lowerCamelCase , param_name="""crop_size""" , default_to_square=__lowerCamelCase ) UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCAmelCase = make_list_of_images(__lowerCamelCase ) if not valid_images(__lowerCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCAmelCase = [convert_to_rgb(__lowerCamelCase ) for image in images] # All transformations expect numpy arrays. UpperCAmelCase = [to_numpy_array(__lowerCamelCase ) for image in images] if do_resize: UpperCAmelCase = [self.resize(image=__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase ) for image in images] if do_center_crop: UpperCAmelCase = [self.center_crop(image=__lowerCamelCase , size=__lowerCamelCase ) for image in images] if do_rescale: UpperCAmelCase = [self.rescale(image=__lowerCamelCase , scale=__lowerCamelCase ) for image in images] if do_normalize: UpperCAmelCase = [self.normalize(image=__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase ) for image in images] UpperCAmelCase = [to_channel_dimension_format(__lowerCamelCase , __lowerCamelCase ) for image in images] UpperCAmelCase = {"""pixel_values""": images} return BatchFeature(data=__lowerCamelCase , tensor_type=__lowerCamelCase )
719
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
0
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __lowercase ( __snake_case , __snake_case ): UpperCamelCase = 1 @register_to_config def __init__( self : Optional[Any] , __lowerCamelCase : int = 1_0_0_0 , __lowerCamelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Union[str, Any]: """simple docstring""" self.set_timesteps(__lowerCamelCase ) # standard deviation of the initial noise distribution UpperCAmelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. UpperCAmelCase = 4 # running values UpperCAmelCase = [] def _lowercase ( self : int , __lowerCamelCase : int , __lowerCamelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" UpperCAmelCase = num_inference_steps UpperCAmelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] UpperCAmelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: UpperCAmelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: UpperCAmelCase = torch.sin(steps * math.pi / 2 ) ** 2 UpperCAmelCase = (1.0 - self.betas**2) ** 0.5 UpperCAmelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] UpperCAmelCase = timesteps.to(__lowerCamelCase ) UpperCAmelCase = [] def _lowercase ( self : Optional[int] , __lowerCamelCase : torch.FloatTensor , __lowerCamelCase : int , __lowerCamelCase : torch.FloatTensor , __lowerCamelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) UpperCAmelCase = (self.timesteps == timestep).nonzero().item() UpperCAmelCase = timestep_index + 1 UpperCAmelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__lowerCamelCase ) if len(self.ets ) == 1: UpperCAmelCase = self.ets[-1] elif len(self.ets ) == 2: UpperCAmelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: UpperCAmelCase = (2_3 * self.ets[-1] - 1_6 * self.ets[-2] + 5 * self.ets[-3]) / 1_2 else: UpperCAmelCase = (1 / 2_4) * (5_5 * self.ets[-1] - 5_9 * self.ets[-2] + 3_7 * self.ets[-3] - 9 * self.ets[-4]) UpperCAmelCase = self._get_prev_sample(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : torch.FloatTensor , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _lowercase ( self : Optional[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.alphas[timestep_index] UpperCAmelCase = self.betas[timestep_index] UpperCAmelCase = self.alphas[prev_timestep_index] UpperCAmelCase = self.betas[prev_timestep_index] UpperCAmelCase = (sample - sigma * ets) / max(__lowerCamelCase , 1e-8 ) UpperCAmelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Union[str, Any] ) -> List[str]: """simple docstring""" return self.config.num_train_timesteps
720
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
0
'''simple docstring''' from __future__ import annotations from collections import Counter from random import random class __lowercase : def __init__( self : Dict ) -> str: """simple docstring""" UpperCAmelCase = {} def _lowercase ( self : Tuple , __lowerCamelCase : str ) -> None: """simple docstring""" UpperCAmelCase = {} def _lowercase ( self : int , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : float ) -> None: """simple docstring""" if nodea not in self.connections: self.add_node(__lowerCamelCase ) if nodea not in self.connections: self.add_node(__lowerCamelCase ) UpperCAmelCase = probability def _lowercase ( self : str ) -> list[str]: """simple docstring""" return list(self.connections ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : str ) -> str: """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->dict[str, int]: UpperCAmelCase = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = Counter(graph.get_nodes() ) UpperCAmelCase = start for _ in range(lowerCAmelCase_ ): UpperCAmelCase = graph.transition(lowerCAmelCase_ ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
721
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
0
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = " " ) ->list: UpperCAmelCase = [] UpperCAmelCase = 0 for index, char in enumerate(lowerCAmelCase_ ): if char == separator: split_words.append(string[last_index:index] ) UpperCAmelCase = index + 1 elif index + 1 == len(lowerCAmelCase_ ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
700
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
701
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
0
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
702
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
0
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __lowercase ( __snake_case ): def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__lowerCamelCase , """tf_padding""" ) ) self.parent.assertTrue(hasattr(__lowerCamelCase , """depth_multiplier""" ) ) class __lowercase : def __init__( self : str , __lowerCamelCase : List[str] , __lowerCamelCase : List[str]=1_3 , __lowerCamelCase : str=3 , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : List[Any]=0.25 , __lowerCamelCase : Union[str, Any]=8 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=1_0_2_4 , __lowerCamelCase : Any=3_2 , __lowerCamelCase : Optional[int]="relu6" , __lowerCamelCase : int=0.1 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Any=True , __lowerCamelCase : int=True , __lowerCamelCase : Union[str, Any]=1_0 , __lowerCamelCase : Union[str, Any]=None , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = image_size UpperCAmelCase = depth_multiplier UpperCAmelCase = min_depth UpperCAmelCase = tf_padding UpperCAmelCase = int(last_hidden_size * depth_multiplier ) UpperCAmelCase = output_stride UpperCAmelCase = hidden_act UpperCAmelCase = classifier_dropout_prob UpperCAmelCase = use_labels UpperCAmelCase = is_training UpperCAmelCase = num_labels UpperCAmelCase = initializer_range UpperCAmelCase = scope def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCAmelCase = self.get_config() return config, pixel_values, labels, pixel_labels def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = MobileNetVaModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _lowercase ( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = MobileNetVaForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () UpperCamelCase = ( {'''feature-extraction''': MobileNetVaModel, '''image-classification''': MobileNetVaForImageClassification} if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = MobileNetVaModelTester(self ) UpperCAmelCase = MobileNetVaConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" pass def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Dict ) -> str: """simple docstring""" def check_hidden_states_output(__lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple ): UpperCAmelCase = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) UpperCAmelCase = outputs.hidden_states UpperCAmelCase = 2_6 self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = MobileNetVaModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->int: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(__lowerCamelCase ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""pt""" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = torch.Size((1, 1_0_0_1) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
703
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
0
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowercase ( __snake_case ): UpperCamelCase = ['''image_processor''', '''tokenizer'''] UpperCamelCase = '''AutoImageProcessor''' UpperCamelCase = '''AutoTokenizer''' def __init__( self : Optional[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ) -> List[str]: """simple docstring""" super().__init__(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self.image_processor def __call__( self : List[Any] , __lowerCamelCase : Any=None , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=None , **__lowerCamelCase : int ) -> Any: """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if images is not None: UpperCAmelCase = self.image_processor(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__lowerCamelCase ) , tensor_type=__lowerCamelCase ) def _lowercase ( self : Any , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Tuple ) -> Tuple: """simple docstring""" return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : int , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : List[str] ) -> Any: """simple docstring""" return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase ) @property def _lowercase ( self : List[str] ) -> Dict: """simple docstring""" return ["input_ids", "attention_mask", "pixel_values"]
704
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
0
'''simple docstring''' import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __a = Mapping[str, np.ndarray] __a = Mapping[str, Any] # Is a nested dict. __a = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class __lowercase : UpperCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. UpperCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. UpperCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. UpperCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. UpperCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions UpperCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files UpperCamelCase = None # Templates used to generate this protein (prediction-only) UpperCamelCase = None # Chain corresponding to each parent UpperCamelCase = None def _UpperCamelCase ( lowerCAmelCase_ ) ->Protein: UpperCAmelCase = R"""(\[[A-Z]+\]\n)""" UpperCAmelCase = [tag.strip() for tag in re.split(lowerCAmelCase_ , lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0] UpperCAmelCase = zip(tags[0::2] , [l.split("""\n""" ) for l in tags[1::2]] ) UpperCAmelCase = ["""N""", """CA""", """C"""] UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None for g in groups: if "[PRIMARY]" == g[0]: UpperCAmelCase = g[1][0].strip() for i in range(len(lowerCAmelCase_ ) ): if seq[i] not in residue_constants.restypes: UpperCAmelCase = """X""" # FIXME: strings are immutable UpperCAmelCase = np.array( [residue_constants.restype_order.get(lowerCAmelCase_ , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: UpperCAmelCase = [] for axis in range(3 ): tertiary.append(list(map(lowerCAmelCase_ , g[1][axis].split() ) ) ) UpperCAmelCase = np.array(lowerCAmelCase_ ) UpperCAmelCase = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(lowerCAmelCase_ ): UpperCAmelCase = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: UpperCAmelCase = np.array(list(map({"""-""": 0, """+""": 1}.get , g[1][0].strip() ) ) ) UpperCAmelCase = np.zeros( ( len(lowerCAmelCase_ ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(lowerCAmelCase_ ): UpperCAmelCase = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=lowerCAmelCase_ , atom_mask=lowerCAmelCase_ , aatype=lowerCAmelCase_ , residue_index=np.arange(len(lowerCAmelCase_ ) ) , b_factors=lowerCAmelCase_ , ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = 0 ) ->List[str]: UpperCAmelCase = [] UpperCAmelCase = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) UpperCAmelCase = prot.parents UpperCAmelCase = prot.parents_chain_index if parents is not None and parents_chain_index is not None: UpperCAmelCase = [p for i, p in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if i == chain_id] if parents is None or len(lowerCAmelCase_ ) == 0: UpperCAmelCase = ["""N/A"""] pdb_headers.append(F"""PARENT {' '.join(lowerCAmelCase_ )}""" ) return pdb_headers def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: UpperCAmelCase = [] UpperCAmelCase = pdb_str.split("""\n""" ) UpperCAmelCase = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) UpperCAmelCase = 4_2 if prot.parents is not None and len(prot.parents ) > 0: UpperCAmelCase = [] if prot.parents_chain_index is not None: UpperCAmelCase = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(lowerCAmelCase_ ) , [] ) parent_dict[str(lowerCAmelCase_ )].append(lowerCAmelCase_ ) UpperCAmelCase = max([int(lowerCAmelCase_ ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): UpperCAmelCase = parent_dict.get(str(lowerCAmelCase_ ) , ["""N/A"""] ) parents_per_chain.append(lowerCAmelCase_ ) else: parents_per_chain.append(list(prot.parents ) ) else: UpperCAmelCase = [["""N/A"""]] def make_parent_line(lowerCAmelCase_ ) -> str: return F"""PARENT {' '.join(lowerCAmelCase_ )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) UpperCAmelCase = 0 for i, l in enumerate(lowerCAmelCase_ ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(lowerCAmelCase_ ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(lowerCAmelCase_ ): UpperCAmelCase = parents_per_chain[chain_counter] else: UpperCAmelCase = ["""N/A"""] out_pdb_lines.append(make_parent_line(lowerCAmelCase_ ) ) return "\n".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = residue_constants.restypes + ["""X"""] def res_atoa(lowerCAmelCase_ ) -> str: return residue_constants.restype_atoa.get(restypes[r] , """UNK""" ) UpperCAmelCase = residue_constants.atom_types UpperCAmelCase = [] UpperCAmelCase = prot.atom_mask UpperCAmelCase = prot.aatype UpperCAmelCase = prot.atom_positions UpperCAmelCase = prot.residue_index.astype(np.intaa ) UpperCAmelCase = prot.b_factors UpperCAmelCase = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("""Invalid aatypes.""" ) UpperCAmelCase = get_pdb_headers(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: pdb_lines.extend(lowerCAmelCase_ ) UpperCAmelCase = aatype.shape[0] UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = string.ascii_uppercase UpperCAmelCase = None # Add all atom sites. for i in range(lowerCAmelCase_ ): UpperCAmelCase = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(lowerCAmelCase_ , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue UpperCAmelCase = """ATOM""" UpperCAmelCase = atom_name if len(lowerCAmelCase_ ) == 4 else F""" {atom_name}""" UpperCAmelCase = """""" UpperCAmelCase = """""" UpperCAmelCase = 1.00 UpperCAmelCase = atom_name[0] # Protein supports only C, N, O, S, this works. UpperCAmelCase = """""" UpperCAmelCase = """A""" if chain_index is not None: UpperCAmelCase = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! UpperCAmelCase = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(lowerCAmelCase_ ) atom_index += 1 UpperCAmelCase = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: UpperCAmelCase = True UpperCAmelCase = chain_index[i + 1] if should_terminate: # Close the chain. UpperCAmelCase = """TER""" UpperCAmelCase = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(lowerCAmelCase_ ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(lowerCAmelCase_ , lowerCAmelCase_ ) ) pdb_lines.append("""END""" ) pdb_lines.append("""""" ) return "\n".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->np.ndarray: return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , ) ->Protein: return Protein( aatype=features["""aatype"""] , atom_positions=result["""final_atom_positions"""] , atom_mask=result["""final_atom_mask"""] , residue_index=features["""residue_index"""] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["""final_atom_mask"""] ) , chain_index=lowerCAmelCase_ , remark=lowerCAmelCase_ , parents=lowerCAmelCase_ , parents_chain_index=lowerCAmelCase_ , )
705
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
0
'''simple docstring''' import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging __a = logging.get_logger(__name__) __a = { """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class __lowercase ( __snake_case ): UpperCamelCase = '''bart''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , __lowerCamelCase : Optional[int]=5_0_2_6_5 , __lowerCamelCase : Optional[Any]=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : List[str]=4_0_9_6 , __lowerCamelCase : str=1_6 , __lowerCamelCase : Dict=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : str=1_6 , __lowerCamelCase : int=0.0 , __lowerCamelCase : str=0.0 , __lowerCamelCase : Tuple="gelu" , __lowerCamelCase : Tuple=1_0_2_4 , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : int=0.0 , __lowerCamelCase : Dict=0.02 , __lowerCamelCase : Any=0.0 , __lowerCamelCase : Any=False , __lowerCamelCase : int=True , __lowerCamelCase : Dict=3 , __lowerCamelCase : int=1 , __lowerCamelCase : Union[str, Any]=0 , __lowerCamelCase : Tuple=2 , __lowerCamelCase : int=True , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Any=2 , **__lowerCamelCase : Union[str, Any] , ) -> Tuple: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = classifier_dropout UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__lowerCamelCase , pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , forced_eos_token_id=__lowerCamelCase , **__lowerCamelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , __lowerCamelCase ): UpperCAmelCase = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" ) class __lowercase ( __snake_case ): @property def _lowercase ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase = {0: """batch"""} UpperCAmelCase = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: UpperCAmelCase = {0: """batch""", 1: """decoder_sequence"""} UpperCAmelCase = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(__lowerCamelCase , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase , UpperCAmelCase = self.num_layers for i in range(__lowerCamelCase ): UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} else: UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def _lowercase ( self : Any ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = super().outputs else: UpperCAmelCase = super(__lowerCamelCase , self ).outputs if self.use_past: UpperCAmelCase , UpperCAmelCase = self.num_layers for i in range(__lowerCamelCase ): UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def _lowercase ( self : Optional[int] , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Generate decoder inputs UpperCAmelCase = seq_length if not self.use_past else 1 UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} UpperCAmelCase = dict(**__lowerCamelCase , **__lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch UpperCAmelCase , UpperCAmelCase = common_inputs["""input_ids"""].shape UpperCAmelCase = common_inputs["""decoder_input_ids"""].shape[1] UpperCAmelCase , UpperCAmelCase = self.num_attention_heads UpperCAmelCase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase = decoder_seq_length + 3 UpperCAmelCase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCAmelCase = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(__lowerCamelCase , __lowerCamelCase )] , dim=1 ) UpperCAmelCase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCAmelCase , UpperCAmelCase = self.num_layers UpperCAmelCase = min(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = max(__lowerCamelCase , __lowerCamelCase ) - min_num_layers UpperCAmelCase = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(__lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase ), ) ) # TODO: test this. UpperCAmelCase = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(__lowerCamelCase , __lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase )) ) return common_inputs def _lowercase ( self : List[Any] , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch UpperCAmelCase , UpperCAmelCase = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values UpperCAmelCase = seqlen + 2 UpperCAmelCase , UpperCAmelCase = self.num_layers UpperCAmelCase , UpperCAmelCase = self.num_attention_heads UpperCAmelCase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase = common_inputs["""attention_mask"""].dtype UpperCAmelCase = torch.cat( [common_inputs["""attention_mask"""], torch.ones(__lowerCamelCase , __lowerCamelCase , dtype=__lowerCamelCase )] , dim=1 ) UpperCAmelCase = [ (torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase )) for _ in range(__lowerCamelCase ) ] return common_inputs def _lowercase ( self : str , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase = compute_effective_axis_dimension( __lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase = tokenizer.num_special_tokens_to_add(__lowerCamelCase ) UpperCAmelCase = compute_effective_axis_dimension( __lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCAmelCase = dict(tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase ) ) return common_inputs def _lowercase ( self : int , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) elif self.task == "causal-lm": UpperCAmelCase = self._generate_dummy_inputs_for_causal_lm( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) else: UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) return common_inputs def _lowercase ( self : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = super()._flatten_past_key_values_(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: UpperCAmelCase = super(__lowerCamelCase , self )._flatten_past_key_values_( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase )
706
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
0
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None ) ->Union[str, Any]: if attention_mask is None: UpperCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_ , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowercase : UpperCamelCase = OPTConfig UpperCamelCase = {} UpperCamelCase = '''gelu''' def __init__( self : Dict , __lowerCamelCase : Any , __lowerCamelCase : Union[str, Any]=1_3 , __lowerCamelCase : int=7 , __lowerCamelCase : List[str]=True , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Any=9_9 , __lowerCamelCase : Optional[int]=1_6 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : str=4 , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : Any=2_0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Dict=1 , __lowerCamelCase : List[Any]=0 , __lowerCamelCase : Union[str, Any]=1_6 , __lowerCamelCase : Union[str, Any]=1_6 , ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = eos_token_id UpperCAmelCase = pad_token_id UpperCAmelCase = bos_token_id UpperCAmelCase = embed_dim UpperCAmelCase = word_embed_proj_dim UpperCAmelCase = False def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__lowerCamelCase , **self.config_updates , ) UpperCAmelCase = prepare_opt_inputs_dict(__lowerCamelCase , __lowerCamelCase ) return config, inputs_dict def _lowercase ( self : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : Dict ) -> int: """simple docstring""" UpperCAmelCase = TFOPTModel(config=__lowerCamelCase ) UpperCAmelCase = inputs_dict["""input_ids"""] UpperCAmelCase = input_ids[:1, :] UpperCAmelCase = inputs_dict["""attention_mask"""][:1, :] UpperCAmelCase = 1 # first forward pass UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase , use_cache=__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase , past_key_values=__lowerCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCAmelCase = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__lowerCamelCase , __lowerCamelCase , rtol=1e-3 ) @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () UpperCamelCase = (TFOPTForCausalLM,) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFOPTModel, '''text-generation''': TFOPTForCausalLM} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = 10 def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = TFOPTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase ) def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__lowerCamelCase ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__lowerCamelCase : int , __lowerCamelCase : Tuple ): if hasattr(__lowerCamelCase , """weight""" ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__lowerCamelCase , """weight""" ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings UpperCAmelCase = model_class(config=__lowerCamelCase ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_input_embeddings() ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__lowerCamelCase ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_input_embeddings() ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. UpperCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __lowerCamelCase ) # check that weights remain the same after resizing UpperCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase = False self.assertTrue(__lowerCamelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __lowerCamelCase ) UpperCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase = False self.assertTrue(__lowerCamelCase ) def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: return tf.constant(lowerCAmelCase_ , dtype=tf.intaa ) @require_tf class __lowercase ( unittest.TestCase ): UpperCamelCase = 99 def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 UpperCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) UpperCAmelCase = input_ids.shape[0] UpperCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase = TFOPTModel.from_pretrained("""facebook/opt-350m""" ) UpperCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) UpperCAmelCase = tf.not_equal(__lowerCamelCase , model.config.pad_token_id ) with tf.GradientTape(): UpperCAmelCase = model(input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ).last_hidden_state UpperCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __lowerCamelCase , atol=4e-3 ) ) UpperCAmelCase = tf.function(__lowerCamelCase , jit_compile=__lowerCamelCase ) UpperCAmelCase = xla_generate(__lowerCamelCase , __lowerCamelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __lowerCamelCase , atol=4e-2 ) ) @require_tf @slow class __lowercase ( unittest.TestCase ): def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" super().setUp() UpperCAmelCase = """facebook/opt-350m""" def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) UpperCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) UpperCAmelCase = [ """Today is a beautiful day and I want to""", """In the city of""", """Paris is the capital of France and""", """Computers and mobile phones have taken""", ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" , padding=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) UpperCAmelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) ) UpperCAmelCase = tf.function(__lowerCamelCase , jit_compile=__lowerCamelCase ) UpperCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) ) @require_tf @slow class __lowercase ( unittest.TestCase ): @property def _lowercase ( self : Any ) -> str: """simple docstring""" return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = """facebook/opt-125m""" UpperCAmelCase = [ """Today is a beautiful day and I want to""", """In the city of New York, the city""", """Paris is the capital of France and the capital""", """Computers and mobile phones have taken over the""", ] UpperCAmelCase = [] UpperCAmelCase = GPTaTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) for prompt in self.prompts: UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(__lowerCamelCase , max_length=1_0 ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Tuple ) -> str: """simple docstring""" UpperCAmelCase = """facebook/opt-350m""" UpperCAmelCase = GPTaTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) UpperCAmelCase = """left""" # use different length sentences to test batching UpperCAmelCase = [ """Hello, my dog is a little""", """Today, I""", ] UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" , padding=__lowerCamelCase ) UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = model.generate(input_ids=__lowerCamelCase , attention_mask=inputs["""attention_mask"""] ) UpperCAmelCase = tokenizer(sentences[0] , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(input_ids=__lowerCamelCase ) UpperCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs["""attention_mask"""][-1] , tf.intaa ) ) UpperCAmelCase = tokenizer(sentences[1] , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(input_ids=__lowerCamelCase , max_length=model.config.max_length - num_paddings ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__lowerCamelCase ) UpperCAmelCase = [ """Hello, my dog is a little bit of a dork.\nI'm a little bit""", """Today, I was in the middle of a conversation with a friend about the""", ] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(__lowerCamelCase , [non_padded_sentence, padded_sentence] ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = """facebook/opt-350m""" UpperCAmelCase = [ """Today is a beautiful day and I want to""", """In the city of San Francisco, the city""", """Paris is the capital of France and the capital""", """Computers and mobile phones have taken over the""", ] UpperCAmelCase = [] UpperCAmelCase = GPTaTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) for prompt in self.prompts: UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(__lowerCamelCase , max_length=1_0 ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
707
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
0
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __lowercase ( __snake_case ): def __init__( self : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple=1_0_2_4 , __lowerCamelCase : Optional[int]=1_0_2_4 , __lowerCamelCase : Optional[Any]=3.6 ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = tokenizer UpperCAmelCase = tokenizer.bos_token_id UpperCAmelCase = dataset UpperCAmelCase = seq_length UpperCAmelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = iter(self.dataset ) UpperCAmelCase = True while more_examples: UpperCAmelCase , UpperCAmelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__lowerCamelCase )["""content"""] ) buffer_len += len(buffer[-1] ) except StopIteration: UpperCAmelCase = False break UpperCAmelCase = tokenizer(__lowerCamelCase , truncation=__lowerCamelCase )["""input_ids"""] UpperCAmelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__lowerCamelCase ) , self.seq_length ): UpperCAmelCase = all_token_ids[i : i + self.seq_length] if len(__lowerCamelCase ) == self.seq_length: yield torch.tensor(__lowerCamelCase ) def _UpperCamelCase ( lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {"""streaming""": True} UpperCAmelCase = load_dataset(args.dataset_name , split="""train""" , **lowerCAmelCase_ ) UpperCAmelCase = ConstantLengthDataset(lowerCAmelCase_ , lowerCAmelCase_ , seq_length=args.seq_length ) UpperCAmelCase = DataLoader(lowerCAmelCase_ , batch_size=args.batch_size ) return eval_dataloader def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[int]: model.eval() UpperCAmelCase = [] for step, batch in enumerate(lowerCAmelCase_ ): with torch.no_grad(): UpperCAmelCase = model(lowerCAmelCase_ , labels=lowerCAmelCase_ ) UpperCAmelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowerCAmelCase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break UpperCAmelCase = torch.mean(torch.cat(lowerCAmelCase_ ) ) try: UpperCAmelCase = torch.exp(lowerCAmelCase_ ) except OverflowError: UpperCAmelCase = float("""inf""" ) return loss.item(), perplexity.item() # Setup Accelerator __a = Accelerator() # Parse configuration __a = HfArgumentParser(EvaluationArguments) __a = parser.parse_args() set_seed(args.seed) # Logging __a = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer __a = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __a = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __a = create_dataloader(args) # Prepare everything with our `accelerator`. __a , __a = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") __a , __a = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
708
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
0
from collections.abc import Sequence from queue import Queue class __lowercase : def __init__( self : Dict , __lowerCamelCase : Dict , __lowerCamelCase : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int=None , __lowerCamelCase : Optional[int]=None ) -> Dict: """simple docstring""" UpperCAmelCase = start UpperCAmelCase = end UpperCAmelCase = val UpperCAmelCase = (start + end) // 2 UpperCAmelCase = left UpperCAmelCase = right def __repr__( self : int ) -> Any: """simple docstring""" return F"""SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})""" class __lowercase : def __init__( self : str , __lowerCamelCase : Sequence , __lowerCamelCase : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = collection UpperCAmelCase = function if self.collection: UpperCAmelCase = self._build_tree(0 , len(__lowerCamelCase ) - 1 ) def _lowercase ( self : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Dict ) -> Union[str, Any]: """simple docstring""" self._update_tree(self.root , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Any ) -> Tuple: """simple docstring""" return self._query_range(self.root , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : int , __lowerCamelCase : str , __lowerCamelCase : List[Any] ) -> List[str]: """simple docstring""" if start == end: return SegmentTreeNode(__lowerCamelCase , __lowerCamelCase , self.collection[start] ) UpperCAmelCase = (start + end) // 2 UpperCAmelCase = self._build_tree(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._build_tree(mid + 1 , __lowerCamelCase ) return SegmentTreeNode(__lowerCamelCase , __lowerCamelCase , self.fn(left.val , right.val ) , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Any , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple ) -> Optional[Any]: """simple docstring""" if node.start == i and node.end == i: UpperCAmelCase = val return if i <= node.mid: self._update_tree(node.left , __lowerCamelCase , __lowerCamelCase ) else: self._update_tree(node.right , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self.fn(node.left.val , node.right.val ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : List[str] , __lowerCamelCase : Optional[int] ) -> Dict: """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , __lowerCamelCase , __lowerCamelCase ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , __lowerCamelCase , node.mid ) , self._query_range(node.right , node.mid + 1 , __lowerCamelCase ) , ) else: # range in right child tree return self._query_range(node.right , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" if self.root is not None: UpperCAmelCase = Queue() queue.put(self.root ) while not queue.empty(): UpperCAmelCase = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print("""*""" * 50) __a = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
709
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
0
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): __a = """pt""" elif is_tf_available(): __a = """tf""" else: __a = """jax""" class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = PerceiverTokenizer UpperCamelCase = False def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" super().setUp() UpperCAmelCase = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _lowercase ( self : Union[str, Any] ) -> int: """simple docstring""" return PerceiverTokenizer.from_pretrained("""deepmind/language-perceiver""" ) def _lowercase ( self : Any , **__lowerCamelCase : List[str] ) -> PerceiverTokenizer: """simple docstring""" return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : List[str] , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Dict=2_0 , __lowerCamelCase : Tuple=5 ) -> Tuple[str, list]: """simple docstring""" UpperCAmelCase = [] for i in range(len(__lowerCamelCase ) ): try: UpperCAmelCase = tokenizer.decode([i] , clean_up_tokenization_spaces=__lowerCamelCase ) except UnicodeDecodeError: pass toks.append((i, tok) ) UpperCAmelCase = list(filter(lambda __lowerCamelCase : re.match(r"""^[ a-zA-Z]+$""" , t[1] ) , __lowerCamelCase ) ) UpperCAmelCase = list(filter(lambda __lowerCamelCase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__lowerCamelCase ) , __lowerCamelCase ) ) if max_length is not None and len(__lowerCamelCase ) > max_length: UpperCAmelCase = toks[:max_length] if min_length is not None and len(__lowerCamelCase ) < min_length and len(__lowerCamelCase ) > 0: while len(__lowerCamelCase ) < min_length: UpperCAmelCase = toks + toks # toks_str = [t[1] for t in toks] UpperCAmelCase = [t[0] for t in toks] # Ensure consistency UpperCAmelCase = tokenizer.decode(__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase ) if " " not in output_txt and len(__lowerCamelCase ) > 1: UpperCAmelCase = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__lowerCamelCase ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__lowerCamelCase ) ) if with_prefix_space: UpperCAmelCase = """ """ + output_txt UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) return output_txt, output_ids def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = self.perceiver_tokenizer UpperCAmelCase = """Unicode €.""" UpperCAmelCase = tokenizer(__lowerCamelCase ) UpperCAmelCase = [4, 9_1, 1_1_6, 1_1_1, 1_0_5, 1_1_7, 1_0_6, 1_0_7, 3_8, 2_3_2, 1_3_6, 1_7_8, 5_2, 5] self.assertEqual(encoded["""input_ids"""] , __lowerCamelCase ) # decoding UpperCAmelCase = tokenizer.decode(__lowerCamelCase ) self.assertEqual(__lowerCamelCase , """[CLS]Unicode €.[SEP]""" ) UpperCAmelCase = tokenizer("""e è é ê ë""" ) UpperCAmelCase = [4, 1_0_7, 3_8, 2_0_1, 1_7_4, 3_8, 2_0_1, 1_7_5, 3_8, 2_0_1, 1_7_6, 3_8, 2_0_1, 1_7_7, 5] self.assertEqual(encoded["""input_ids"""] , __lowerCamelCase ) # decoding UpperCAmelCase = tokenizer.decode(__lowerCamelCase ) self.assertEqual(__lowerCamelCase , """[CLS]e è é ê ë[SEP]""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """[CLS]e è é ê ë[SEP]""" ) def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.perceiver_tokenizer UpperCAmelCase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off UpperCAmelCase = [4, 7_1, 3_8, 1_1_4, 1_1_7, 1_1_6, 1_0_9, 3_8, 1_1_8, 1_0_3, 1_2_0, 1_0_3, 1_0_9, 1_2_0, 1_0_3, 1_1_8, 1_1_0, 3_8, 1_0_8, 1_1_7, 1_2_0, 3_8, 1_2_1, 1_2_3, 1_1_5, 1_1_5, 1_0_3, 1_2_0, 1_1_1, 1_2_8, 1_0_3, 1_2_2, 1_1_1, 1_1_7, 1_1_6, 5_2, 5, 0] # fmt: on UpperCAmelCase = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors=__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) if FRAMEWORK != "jax": UpperCAmelCase = list(batch.input_ids.numpy()[0] ) else: UpperCAmelCase = list(batch.input_ids.tolist()[0] ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertEqual((2, 3_8) , batch.input_ids.shape ) self.assertEqual((2, 3_8) , batch.attention_mask.shape ) def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.perceiver_tokenizer UpperCAmelCase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] UpperCAmelCase = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors=__lowerCamelCase ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , __lowerCamelCase ) self.assertIn("""attention_mask""" , __lowerCamelCase ) self.assertNotIn("""decoder_input_ids""" , __lowerCamelCase ) self.assertNotIn("""decoder_attention_mask""" , __lowerCamelCase ) def _lowercase ( self : Tuple ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.perceiver_tokenizer UpperCAmelCase = [ """Summary of the text.""", """Another summary.""", ] UpperCAmelCase = tokenizer( text_target=__lowerCamelCase , max_length=3_2 , padding="""max_length""" , truncation=__lowerCamelCase , return_tensors=__lowerCamelCase ) self.assertEqual(3_2 , targets["""input_ids"""].shape[1] ) def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 4_2 ) # Now let's start the test UpperCAmelCase = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = """ He is very happy, UNwant\u00E9d,running""" UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) tokenizer.save_pretrained(__lowerCamelCase ) UpperCAmelCase = tokenizer.__class__.from_pretrained(__lowerCamelCase ) UpperCAmelCase = after_tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) shutil.rmtree(__lowerCamelCase ) UpperCAmelCase = self.get_tokenizers(model_max_length=4_2 ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) UpperCAmelCase = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) tokenizer.save_pretrained(__lowerCamelCase ) UpperCAmelCase = tokenizer.__class__.from_pretrained(__lowerCamelCase ) UpperCAmelCase = after_tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 4_2 ) UpperCAmelCase = tokenizer.__class__.from_pretrained(__lowerCamelCase , model_max_length=4_3 ) self.assertEqual(tokenizer.model_max_length , 4_3 ) shutil.rmtree(__lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowerCamelCase ) with open(os.path.join(__lowerCamelCase , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: UpperCAmelCase = json.load(__lowerCamelCase ) with open(os.path.join(__lowerCamelCase , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: UpperCAmelCase = json.load(__lowerCamelCase ) UpperCAmelCase = [F"""<extra_id_{i}>""" for i in range(1_2_5 )] UpperCAmelCase = added_tokens_extra_ids + [ """an_additional_special_token""" ] UpperCAmelCase = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(__lowerCamelCase , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(__lowerCamelCase , __lowerCamelCase ) with open(os.path.join(__lowerCamelCase , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(__lowerCamelCase , __lowerCamelCase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files UpperCAmelCase = tokenizer_class.from_pretrained( __lowerCamelCase , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained UpperCAmelCase = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=__lowerCamelCase )] UpperCAmelCase = tokenizer_class.from_pretrained( __lowerCamelCase , additional_special_tokens=__lowerCamelCase , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_7_8] ) , """�""" ) def _lowercase ( self : Optional[int] ) -> int: """simple docstring""" pass def _lowercase ( self : Tuple ) -> Any: """simple docstring""" pass def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" pass def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" pass def _lowercase ( self : int ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizers(fast=__lowerCamelCase , do_lower_case=__lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): UpperCAmelCase = ["""[CLS]""", """t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """s""", """t""", """[SEP]"""] UpperCAmelCase = tokenizer.convert_tokens_to_string(__lowerCamelCase ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase )
710
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = { """configuration_clipseg""": [ """CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CLIPSegConfig""", """CLIPSegTextConfig""", """CLIPSegVisionConfig""", ], """processing_clipseg""": ["""CLIPSegProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST""", """CLIPSegModel""", """CLIPSegPreTrainedModel""", """CLIPSegTextModel""", """CLIPSegVisionModel""", """CLIPSegForImageSegmentation""", ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
711
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = { """configuration_xlm_roberta_xl""": [ """XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaXLConfig""", """XLMRobertaXLOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaXLForCausalLM""", """XLMRobertaXLForMaskedLM""", """XLMRobertaXLForMultipleChoice""", """XLMRobertaXLForQuestionAnswering""", """XLMRobertaXLForSequenceClassification""", """XLMRobertaXLForTokenClassification""", """XLMRobertaXLModel""", """XLMRobertaXLPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
712
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
0