code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = {'openai-gpt': 'https://huggingface.co/openai-gpt/resolve/main/config.json'} class __lowercase ( __lowerCamelCase ): snake_case_ = """openai-gpt""" snake_case_ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : int ,A : Any=40_478 ,A : List[str]=512 ,A : Tuple=768 ,A : int=12 ,A : Union[str, Any]=12 ,A : int="gelu" ,A : List[str]=0.1 ,A : int=0.1 ,A : List[Any]=0.1 ,A : Optional[int]=1e-5 ,A : Optional[Any]=0.0_2 ,A : List[Any]="cls_index" ,A : List[str]=True ,A : Union[str, Any]=None ,A : int=True ,A : int=0.1 ,**A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : List[str] = vocab_size UpperCAmelCase__ : Optional[int] = n_positions UpperCAmelCase__ : Dict = n_embd UpperCAmelCase__ : List[str] = n_layer UpperCAmelCase__ : List[Any] = n_head UpperCAmelCase__ : str = afn UpperCAmelCase__ : Dict = resid_pdrop UpperCAmelCase__ : Union[str, Any] = embd_pdrop UpperCAmelCase__ : Optional[Any] = attn_pdrop UpperCAmelCase__ : Tuple = layer_norm_epsilon UpperCAmelCase__ : Any = initializer_range UpperCAmelCase__ : Union[str, Any] = summary_type UpperCAmelCase__ : Tuple = summary_use_proj UpperCAmelCase__ : List[str] = summary_activation UpperCAmelCase__ : Optional[Any] = summary_first_dropout UpperCAmelCase__ : str = summary_proj_to_labels super().__init__(**A )
65
"""simple docstring""" __UpperCAmelCase = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image', 'mask_image']) __UpperCAmelCase = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['example_image', 'image', 'mask_image']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset(['input_tokens']) __UpperCAmelCase = frozenset(['input_tokens'])
65
1
"""simple docstring""" __UpperCAmelCase = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image', 'mask_image']) __UpperCAmelCase = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['example_image', 'image', 'mask_image']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset(['input_tokens']) __UpperCAmelCase = frozenset(['input_tokens'])
65
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase__ : Tuple = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[Any] = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase__ : int = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 16_000, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase__ : Optional[int] = tempfile.mkdtemp() UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,A ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.feature_extraction_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) # load decoder from hub UpperCAmelCase__ : int = """hf-internal-testing/ngram-beam-search-decoder""" def __lowercase ( self : str ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.add_kwargs_tokens_map.copy() kwargs.update(A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[str] ,**A : Dict ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Any ,**A : List[Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name ,**A ) def __lowercase ( self : Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_decoder() UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() ,feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor ,A ) # decoder self.assertEqual(processor.decoder._alphabet.labels ,decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set ,decoder.model_container[decoder._model_key]._unigram_set ,) self.assertIsInstance(processor.decoder ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname ,alpha=5.0 ,beta=3.0 ,score_boundary=-7.0 ,unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha ,5.0 ) self.assertEqual(processor.language_model.beta ,3.0 ) self.assertEqual(processor.language_model.score_boundary ,-7.0 ) self.assertEqual(processor.language_model.unk_score_offset ,3 ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(A ,"""include""" ): WavaVecaProcessorWithLM( tokenizer=A ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : Optional[Any] = feature_extractor(A ,return_tensors="""np""" ) UpperCAmelCase__ : List[Any] = processor(A ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : List[Any] = """This is a test string""" UpperCAmelCase__ : int = processor(text=A ) UpperCAmelCase__ : Dict = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __lowercase ( self : Tuple ,A : List[Any]=(2, 10, 16) ,A : Dict=77 ): '''simple docstring''' np.random.seed(A ) return np.random.rand(*A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits(shape=(10, 16) ,seed=13 ) UpperCAmelCase__ : Tuple = processor.decode(A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(A )[0] self.assertEqual(decoded_decoder[0] ,decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" ,decoded_processor.text ) self.assertEqual(decoded_decoder[-2] ,decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] ,decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def __lowercase ( self : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_feature_extractor() UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : List[str] = processor.batch_decode(A ) else: with get_context(A ).Pool() as pool: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(A ,A ) UpperCAmelCase__ : Optional[Any] = list(A ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A ,decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] ,decoded_processor.text ) self.assertListEqual(A ,decoded_processor.logit_score ) self.assertListEqual(A ,decoded_processor.lm_score ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits() UpperCAmelCase__ : Any = 15 UpperCAmelCase__ : Dict = -2_0.0 UpperCAmelCase__ : List[Any] = -4.0 UpperCAmelCase__ : Union[str, Any] = processor.batch_decode( A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[str] = decoded_processor_out.text UpperCAmelCase__ : List[str] = list(A ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( A ,A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] ,A ) self.assertTrue(np.array_equal(A ,decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] ,A ,atol=1e-3 ) ) self.assertTrue(np.array_equal(A ,decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] ,A ,atol=1e-3 ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits() UpperCAmelCase__ : Tuple = 2.0 UpperCAmelCase__ : str = 5.0 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = processor.batch_decode( A ,alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) UpperCAmelCase__ : Any = decoded_processor_out.text UpperCAmelCase__ : Union[str, Any] = list(A ) decoder.reset_params( alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch( A ,A ,) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] ,A ) UpperCAmelCase__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha ,2.0 ) self.assertEqual(lm_model.beta ,5.0 ) self.assertEqual(lm_model.unk_score_offset ,-2_0.0 ) self.assertEqual(lm_model.score_boundary ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : str = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Any = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[int] = os.listdir(A ) UpperCAmelCase__ : List[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(A ) UpperCAmelCase__ : Tuple = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[int] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Tuple = os.listdir(A ) UpperCAmelCase__ : Dict = os.listdir(A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = floats_list((3, 1_000) ) UpperCAmelCase__ : List[str] = processor_wavaveca(A ,return_tensors="""np""" ) UpperCAmelCase__ : Dict = processor_auto(A ,return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() ,input_auto[key].sum() ,delta=1e-2 ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Tuple = processor_wavaveca.batch_decode(A ) UpperCAmelCase__ : List[str] = processor_auto.batch_decode(A ) self.assertListEqual(decoded_wavaveca.text ,decoded_auto.text ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : int = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) self.assertListEqual( processor.model_input_names ,feature_extractor.model_input_names ,msg="""`processor` and `feature_extractor` model input names do not match""" ,) @staticmethod def __lowercase ( A : Optional[Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = self._get_dummy_logits()[0] UpperCAmelCase__ : List[str] = processor.decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ) ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""end_offset""" ) ,[1, 3, 5] ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : int = self._get_dummy_logits() UpperCAmelCase__ : Any = processor.batch_decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(A ,"""word""" ) ) for o in outputs["""word_offsets"""]] ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""end_offset""" ) ,[1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self : Tuple ): '''simple docstring''' import torch UpperCAmelCase__ : Any = load_dataset("""common_voice""" ,"""en""" ,split="""train""" ,streaming=A ) UpperCAmelCase__ : Tuple = ds.cast_column("""audio""" ,datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Tuple = iter(A ) UpperCAmelCase__ : Optional[int] = next(A ) UpperCAmelCase__ : List[Any] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase__ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Tuple = processor(sample["""audio"""]["""array"""] ,return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(A ).logits.cpu().numpy() UpperCAmelCase__ : Any = processor.decode(logits[0] ,output_word_offsets=A ) UpperCAmelCase__ : str = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase__ : Dict = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,A ) self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,output.text ) # output times UpperCAmelCase__ : str = torch.tensor(self.get_from_offsets(A ,"""start_time""" ) ) UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(A ,"""end_time""" ) ) # fmt: off UpperCAmelCase__ : Union[str, Any] = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : List[Any] = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) ) self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) )
65
1
"""simple docstring""" from __future__ import annotations from collections.abc import Generator def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : dict[int, int] = {} UpperCAmelCase__ : Optional[int] = 2 while True: UpperCAmelCase__ : Dict = factor_map.pop(__UpperCamelCase , __UpperCamelCase ) if factor: UpperCAmelCase__ : Tuple = factor + prime while x in factor_map: x += factor UpperCAmelCase__ : int = factor else: UpperCAmelCase__ : Optional[Any] = prime yield prime prime += 1 def lowerCAmelCase ( __UpperCamelCase = 1e10 ): '''simple docstring''' UpperCAmelCase__ : Any = sieve() UpperCAmelCase__ : Any = 1 while True: UpperCAmelCase__ : List[Any] = next(__UpperCamelCase ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(__UpperCamelCase ) n += 2 if __name__ == "__main__": print(solution())
65
"""simple docstring""" from sklearn.metrics import fa_score import datasets __UpperCAmelCase = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __UpperCAmelCase = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __UpperCAmelCase = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def __lowercase ( self : List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) ,reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"""] ,) def __lowercase ( self : Union[str, Any] ,A : List[str] ,A : List[Any] ,A : Optional[Any]=None ,A : List[str]=1 ,A : Optional[Any]="binary" ,A : Any=None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = fa_score( A ,A ,labels=A ,pos_label=A ,average=A ,sample_weight=A ) return {"f1": float(A ) if score.size == 1 else score}
65
1
"""simple docstring""" import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( enum.Enum ): snake_case_ = 0 snake_case_ = 1 @add_end_docstrings(__lowerCamelCase ) class __lowercase ( __lowerCamelCase ): snake_case_ = """generated""" def __init__( self : List[str] ,*A : Tuple ,**A : str ): '''simple docstring''' super().__init__(*A ,**A ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == """tf""" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def __lowercase ( self : Optional[Any] ,A : List[Any]=None ,A : Any=None ,A : Optional[int]=None ,A : Union[str, Any]=None ,A : Optional[Any]=None ,A : Optional[Any]=None ,**A : str ,): '''simple docstring''' UpperCAmelCase__ : Any = {} if truncation is not None: UpperCAmelCase__ : Any = truncation UpperCAmelCase__ : Union[str, Any] = generate_kwargs UpperCAmelCase__ : Dict = {} if return_tensors is not None and return_type is None: UpperCAmelCase__ : Optional[Any] = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: UpperCAmelCase__ : Optional[int] = return_type if clean_up_tokenization_spaces is not None: UpperCAmelCase__ : Optional[int] = clean_up_tokenization_spaces if stop_sequence is not None: UpperCAmelCase__ : List[Any] = self.tokenizer.encode(A ,add_special_tokens=A ) if len(A ) > 1: warnings.warn( """Stopping on a multiple token sequence is not yet supported on transformers. The first token of""" """ the stop sequence will be used as the stop sequence string in the interim.""" ) UpperCAmelCase__ : Union[str, Any] = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __lowercase ( self : Dict ,A : int ,A : int ,A : int ): '''simple docstring''' return True def __lowercase ( self : List[Any] ,*A : Dict ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : str = self.model.config.prefix if self.model.config.prefix is not None else """""" if isinstance(args[0] ,A ): if self.tokenizer.pad_token_id is None: raise ValueError("""Please make sure that the tokenizer has a pad_token_id when using a batch input""" ) UpperCAmelCase__ : Union[str, Any] = ([prefix + arg for arg in args[0]],) UpperCAmelCase__ : int = True elif isinstance(args[0] ,A ): UpperCAmelCase__ : Optional[Any] = (prefix + args[0],) UpperCAmelCase__ : List[Any] = False else: raise ValueError( f" `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`" ) UpperCAmelCase__ : Dict = self.tokenizer(*A ,padding=A ,truncation=A ,return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Dict ,*A : str ,**A : str ): '''simple docstring''' UpperCAmelCase__ : str = super().__call__(*A ,**A ) if ( isinstance(args[0] ,A ) and all(isinstance(A ,A ) for el in args[0] ) and all(len(A ) == 1 for res in result ) ): return [res[0] for res in result] return result def __lowercase ( self : Dict ,A : Any ,A : str=TruncationStrategy.DO_NOT_TRUNCATE ,**A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self._parse_and_tokenize(A ,truncation=A ,**A ) return inputs def __lowercase ( self : Any ,A : List[str] ,**A : Optional[Any] ): '''simple docstring''' if self.framework == "pt": UpperCAmelCase__ , UpperCAmelCase__ : Tuple = model_inputs["""input_ids"""].shape elif self.framework == "tf": UpperCAmelCase__ , UpperCAmelCase__ : Dict = tf.shape(model_inputs["""input_ids"""] ).numpy() UpperCAmelCase__ : str = generate_kwargs.get("""min_length""" ,self.model.config.min_length ) UpperCAmelCase__ : str = generate_kwargs.get("""max_length""" ,self.model.config.max_length ) self.check_inputs(A ,generate_kwargs["""min_length"""] ,generate_kwargs["""max_length"""] ) UpperCAmelCase__ : Dict = self.model.generate(**A ,**A ) UpperCAmelCase__ : Dict = output_ids.shape[0] if self.framework == "pt": UpperCAmelCase__ : List[str] = output_ids.reshape(A ,out_b // in_b ,*output_ids.shape[1:] ) elif self.framework == "tf": UpperCAmelCase__ : int = tf.reshape(A ,(in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def __lowercase ( self : List[Any] ,A : Any ,A : str=ReturnType.TEXT ,A : Any=False ): '''simple docstring''' UpperCAmelCase__ : Any = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: UpperCAmelCase__ : Optional[Any] = {f"{self.return_name}_token_ids": output_ids} elif return_type == ReturnType.TEXT: UpperCAmelCase__ : str = { f"{self.return_name}_text": self.tokenizer.decode( A ,skip_special_tokens=A ,clean_up_tokenization_spaces=A ,) } records.append(A ) return records @add_end_docstrings(__lowerCamelCase ) class __lowercase ( __lowerCamelCase ): snake_case_ = """summary""" def __call__( self : Optional[Any] ,*A : Any ,**A : Optional[Any] ): '''simple docstring''' return super().__call__(*A ,**A ) def __lowercase ( self : Union[str, Any] ,A : int ,A : int ,A : int ): '''simple docstring''' if max_length < min_length: logger.warning(f"Your min_length={min_length} must be inferior than your max_length={max_length}." ) if input_length < max_length: logger.warning( f"Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is " """a summarization task, where outputs shorter than the input are typically wanted, you might """ f"consider decreasing max_length manually, e.g. summarizer('...', max_length={input_length//2})" ) @add_end_docstrings(__lowerCamelCase ) class __lowercase ( __lowerCamelCase ): snake_case_ = """translation""" def __lowercase ( self : List[str] ,A : int ,A : int ,A : int ): '''simple docstring''' if input_length > 0.9 * max_length: logger.warning( f"Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider " """increasing your max_length manually, e.g. translator('...', max_length=400)""" ) return True def __lowercase ( self : int ,*A : Any ,A : Tuple=TruncationStrategy.DO_NOT_TRUNCATE ,A : str=None ,A : Optional[int]=None ): '''simple docstring''' if getattr(self.tokenizer ,"""_build_translation_inputs""" ,A ): return self.tokenizer._build_translation_inputs( *A ,return_tensors=self.framework ,truncation=A ,src_lang=A ,tgt_lang=A ) else: return super()._parse_and_tokenize(*A ,truncation=A ) def __lowercase ( self : Union[str, Any] ,A : Dict=None ,A : Union[str, Any]=None ,**A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = super()._sanitize_parameters(**A ) if src_lang is not None: UpperCAmelCase__ : List[str] = src_lang if tgt_lang is not None: UpperCAmelCase__ : List[Any] = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. UpperCAmelCase__ : Optional[Any] = kwargs.get("""task""" ,self.task ) UpperCAmelCase__ : Any = task.split("""_""" ) if task and len(A ) == 4: # translation, XX, to YY UpperCAmelCase__ : str = items[1] UpperCAmelCase__ : Union[str, Any] = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] ,*A : int ,**A : Union[str, Any] ): '''simple docstring''' return super().__call__(*A ,**A )
65
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece.model') __UpperCAmelCase = {'target_lang': 'fi', 'source_lang': 'en'} __UpperCAmelCase = '>>zh<<' __UpperCAmelCase = 'Helsinki-NLP/' if is_torch_available(): __UpperCAmelCase = 'pt' elif is_tf_available(): __UpperCAmelCase = 'tf' else: __UpperCAmelCase = 'jax' @require_sentencepiece class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = MarianTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Optional[int] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Optional[Any] = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] UpperCAmelCase__ : int = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[int] = Path(self.tmpdirname ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(A ,save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(A ,save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) UpperCAmelCase__ : Dict = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self : List[Any] ,**A : List[Any] ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Union[str, Any] ,A : Tuple ): '''simple docstring''' return ( "This is a test", "This is a test", ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """</s>""" UpperCAmelCase__ : int = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""</s>""" ) self.assertEqual(vocab_keys[1] ,"""<unk>""" ) self.assertEqual(vocab_keys[-1] ,"""<pad>""" ) self.assertEqual(len(A ) ,9 ) def __lowercase ( self : Dict ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,9 ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MarianTokenizer.from_pretrained(f"{ORG_NAME}opus-mt-en-de" ) UpperCAmelCase__ : List[str] = en_de_tokenizer(["""I am a small frog"""] ,return_tensors=A ) self.assertIsInstance(A ,A ) UpperCAmelCase__ : str = [38, 121, 14, 697, 38_848, 0] self.assertListEqual(A ,batch.input_ids[0] ) UpperCAmelCase__ : Optional[Any] = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(A ) UpperCAmelCase__ : Tuple = [x.name for x in Path(A ).glob("""*""" )] self.assertIn("""source.spm""" ,A ) MarianTokenizer.from_pretrained(A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = tok( ["""I am a small frog""" * 1_000, """I am a small frog"""] ,padding=A ,truncation=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch.input_ids.shape ,(2, 512) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : Tuple = tok(["""I am a tiny frog""", """I am a small frog"""] ,padding=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch_smaller.input_ids.shape ,(2, 10) ) @slow def __lowercase ( self : Dict ): '''simple docstring''' # fmt: off UpperCAmelCase__ : Optional[int] = {"""input_ids""": [[43_495, 462, 20, 42_164, 1_369, 52, 464, 132, 1_703, 492, 13, 7_491, 38_999, 6, 8, 464, 132, 1_703, 492, 13, 4_669, 37_867, 13, 7_525, 27, 1_593, 988, 13, 33_972, 7_029, 6, 20, 8_251, 383, 2, 270, 5_866, 3_788, 2, 2_353, 8_251, 12_338, 2, 13_958, 387, 2, 3_629, 6_953, 188, 2_900, 2, 13_958, 8_011, 11_501, 23, 8_460, 4_073, 34_009, 20, 435, 11_439, 27, 8, 8_460, 4_073, 6_004, 20, 9_988, 375, 27, 33, 266, 1_945, 1_076, 1_350, 37_867, 3_288, 5, 577, 1_076, 4_374, 8, 5_082, 5, 26_453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10_767, 6, 316, 304, 4_239, 3, 0], [148, 15_722, 19, 1_839, 12, 1_350, 13, 22_327, 5_082, 5_418, 47_567, 35_938, 59, 318, 19_552, 108, 2_183, 54, 14_976, 4_835, 32, 547, 1_114, 8, 315, 2_417, 5, 92, 19_088, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100], [36, 6_395, 12_570, 39_147, 11_597, 6, 266, 4, 45_405, 7_296, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""Helsinki-NLP/opus-mt-en-de""" ,revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" ,decode_kwargs={"""use_source_tokenizer""": True} ,) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) UpperCAmelCase__ : Any = """Tämä on testi""" UpperCAmelCase__ : int = """This is a test""" UpperCAmelCase__ : List[str] = [76, 7, 2_047, 2] UpperCAmelCase__ : Optional[Any] = [69, 12, 11, 940, 2] UpperCAmelCase__ : List[str] = tokenizer(A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : Optional[int] = tokenizer(text_target=A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : int = tokenizer.decode(A ,skip_special_tokens=A ) self.assertEqual(A ,A )
65
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPanoramaPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() @skip_mps class __lowercase ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): snake_case_ = StableDiffusionPanoramaPipeline snake_case_ = TEXT_TO_IMAGE_PARAMS snake_case_ = TEXT_TO_IMAGE_BATCH_PARAMS snake_case_ = TEXT_TO_IMAGE_IMAGE_PARAMS snake_case_ = TEXT_TO_IMAGE_IMAGE_PARAMS def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=1 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") ,up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") ,cross_attention_dim=32 ,) UpperCAmelCase__ : int = DDIMScheduler() torch.manual_seed(0 ) UpperCAmelCase__ : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=4 ,) torch.manual_seed(0 ) UpperCAmelCase__ : Optional[int] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_000 ,) UpperCAmelCase__ : Optional[Any] = CLIPTextModel(A ) UpperCAmelCase__ : List[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCAmelCase__ : List[str] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def __lowercase ( self : Union[str, Any] ,A : Dict ,A : Tuple=0 ): '''simple docstring''' UpperCAmelCase__ : List[str] = torch.manual_seed(A ) UpperCAmelCase__ : Union[str, Any] = { """prompt""": """a photo of the dolomites""", """generator""": generator, # Setting height and width to None to prevent OOMs on CPU. """height""": None, """width""": None, """num_inference_steps""": 1, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ : str = self.get_dummy_components() UpperCAmelCase__ : Any = StableDiffusionPanoramaPipeline(**A ) UpperCAmelCase__ : Dict = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(A ) UpperCAmelCase__ : Dict = sd_pipe(**A ).images UpperCAmelCase__ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Union[str, Any] = np.array([0.6_1_8_6, 0.5_3_7_4, 0.4_9_1_5, 0.4_1_3_5, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_7, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __lowercase ( self : List[str] ): '''simple docstring''' super().test_inference_batch_consistent(batch_sizes=[1, 2] ) def __lowercase ( self : List[Any] ): '''simple docstring''' super().test_inference_batch_single_identical(batch_size=2 ,expected_max_diff=3.25e-3 ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ : str = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = StableDiffusionPanoramaPipeline(**A ) UpperCAmelCase__ : Optional[Any] = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : List[Any] = self.get_dummy_inputs(A ) UpperCAmelCase__ : Optional[int] = """french fries""" UpperCAmelCase__ : List[str] = sd_pipe(**A ,negative_prompt=A ) UpperCAmelCase__ : Optional[Any] = output.images UpperCAmelCase__ : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : str = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ : str = self.get_dummy_components() UpperCAmelCase__ : Dict = StableDiffusionPanoramaPipeline(**A ) UpperCAmelCase__ : List[str] = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Any = self.get_dummy_inputs(A ) UpperCAmelCase__ : List[str] = sd_pipe(**A ,view_batch_size=2 ) UpperCAmelCase__ : Dict = output.images UpperCAmelCase__ : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Optional[Any] = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : str = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ : Optional[int] = self.get_dummy_components() UpperCAmelCase__ : List[str] = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule="""scaled_linear""" ) UpperCAmelCase__ : Optional[Any] = StableDiffusionPanoramaPipeline(**A ) UpperCAmelCase__ : List[str] = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = self.get_dummy_inputs(A ) UpperCAmelCase__ : Tuple = sd_pipe(**A ).images UpperCAmelCase__ : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Any = np.array([0.4_0_2_4, 0.6_5_1_0, 0.4_9_0_1, 0.5_3_7_8, 0.5_8_1_3, 0.5_6_2_2, 0.4_7_9_5, 0.4_4_6_7, 0.4_9_5_2] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : str = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ : List[Any] = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = PNDMScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule="""scaled_linear""" ,skip_prk_steps=A ) UpperCAmelCase__ : List[str] = StableDiffusionPanoramaPipeline(**A ) UpperCAmelCase__ : int = sd_pipe.to(A ) sd_pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : List[str] = self.get_dummy_inputs(A ) UpperCAmelCase__ : int = sd_pipe(**A ).images UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : str = np.array([0.6_3_9_1, 0.6_2_9_1, 0.4_8_6_1, 0.5_1_3_4, 0.5_5_5_2, 0.4_5_7_8, 0.5_0_3_2, 0.5_0_2_3, 0.4_5_3_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : List[str] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : int ,A : Dict=0 ): '''simple docstring''' UpperCAmelCase__ : Dict = torch.manual_seed(A ) UpperCAmelCase__ : Tuple = { """prompt""": """a photo of the dolomites""", """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """stabilityai/stable-diffusion-2-base""" UpperCAmelCase__ : Union[str, Any] = DDIMScheduler.from_pretrained(A ,subfolder="""scheduler""" ) UpperCAmelCase__ : List[str] = StableDiffusionPanoramaPipeline.from_pretrained(A ,scheduler=A ,safety_checker=A ) pipe.to(A ) pipe.set_progress_bar_config(disable=A ) pipe.enable_attention_slicing() UpperCAmelCase__ : Union[str, Any] = self.get_inputs() UpperCAmelCase__ : Tuple = pipe(**A ).images UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 2_048, 3) UpperCAmelCase__ : List[Any] = np.array( [ 0.3_6_9_6_8_3_9_2, 0.2_7_0_2_5_3_7_2, 0.3_2_4_4_6_7_6_6, 0.2_8_3_7_9_3_8_7, 0.3_6_3_6_3_2_7_4, 0.3_0_7_3_3_3_4_7, 0.2_7_1_0_0_0_2_7, 0.2_7_0_5_4_1_2_5, 0.2_5_5_3_6_0_9_6, ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-2 def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : str = StableDiffusionPanoramaPipeline.from_pretrained( """stabilityai/stable-diffusion-2-base""" ,safety_checker=A ) UpperCAmelCase__ : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(A ) pipe.set_progress_bar_config(disable=A ) pipe.enable_attention_slicing() UpperCAmelCase__ : List[str] = self.get_inputs() UpperCAmelCase__ : Union[str, Any] = pipe(**A ).images UpperCAmelCase__ : List[str] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 2_048, 3) UpperCAmelCase__ : Any = np.array( [ [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ] ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = 0 def callback_fn(A : int ,A : int ,A : torch.FloatTensor ) -> None: UpperCAmelCase__ : Union[str, Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: UpperCAmelCase__ : Tuple = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 256) UpperCAmelCase__ : Union[str, Any] = latents[0, -3:, -3:, -1] UpperCAmelCase__ : Any = np.array( [ 0.1_8_6_8_1_8_6_9, 0.3_3_9_0_7_8_1_6, 0.5_3_6_1_2_7_6, 0.1_4_4_3_2_8_6_5, -0.0_2_8_5_6_6_1_1, -0.7_3_9_4_1_1_2_3, 0.2_3_3_9_7_9_8_7, 0.4_7_3_2_2_6_8_2, -0.3_7_8_2_3_1_6_4, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: UpperCAmelCase__ : Any = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 256) UpperCAmelCase__ : Union[str, Any] = latents[0, -3:, -3:, -1] UpperCAmelCase__ : int = np.array( [ 0.1_8_5_3_9_6_4_5, 0.3_3_9_8_7_2_4_8, 0.5_3_7_8_5_5_9, 0.1_4_4_3_7_1_4_2, -0.0_2_4_5_5_2_6_1, -0.7_3_3_8_3_1_7, 0.2_3_9_9_0_7_5_5, 0.4_7_3_5_6_2_7_2, -0.3_7_8_6_5_0_5, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 UpperCAmelCase__ : str = False UpperCAmelCase__ : Union[str, Any] = """stabilityai/stable-diffusion-2-base""" UpperCAmelCase__ : Optional[int] = DDIMScheduler.from_pretrained(A ,subfolder="""scheduler""" ) UpperCAmelCase__ : int = StableDiffusionPanoramaPipeline.from_pretrained(A ,scheduler=A ,safety_checker=A ) UpperCAmelCase__ : List[Any] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) pipe.enable_attention_slicing() UpperCAmelCase__ : List[str] = self.get_inputs() pipe(**A ,callback=A ,callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def __lowercase ( self : List[Any] ): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCAmelCase__ : int = """stabilityai/stable-diffusion-2-base""" UpperCAmelCase__ : List[str] = DDIMScheduler.from_pretrained(A ,subfolder="""scheduler""" ) UpperCAmelCase__ : Dict = StableDiffusionPanoramaPipeline.from_pretrained(A ,scheduler=A ,safety_checker=A ) UpperCAmelCase__ : List[str] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() UpperCAmelCase__ : str = self.get_inputs() UpperCAmelCase__ : Union[str, Any] = pipe(**A ) UpperCAmelCase__ : int = torch.cuda.max_memory_allocated() # make sure that less than 5.2 GB is allocated assert mem_bytes < 5.5 * 10**9
65
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=__lowerCamelCase ): snake_case_ = ["""onnx"""] def __init__( self : int ,*A : List[str] ,**A : int ): '''simple docstring''' requires_backends(self ,["""onnx"""] ) @classmethod def __lowercase ( cls : Optional[Any] ,*A : List[str] ,**A : Dict ): '''simple docstring''' requires_backends(cls ,["""onnx"""] ) @classmethod def __lowercase ( cls : List[Any] ,*A : Optional[int] ,**A : int ): '''simple docstring''' requires_backends(cls ,["""onnx"""] )
65
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCAmelCase = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) UpperCAmelCase__ : List[Any] = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(__UpperCamelCase ) # Let's go UpperCAmelCase__ : int = parser.parse_args() if not hasattr(__UpperCamelCase , """func""" ): parser.print_help() exit(1 ) # Run UpperCAmelCase__ : Union[str, Any] = args.func(__UpperCamelCase ) service.run() if __name__ == "__main__": main()
65
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __UpperCAmelCase = { 'configuration_longt5': ['LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LongT5Config', 'LongT5OnnxConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'LongT5EncoderModel', 'LongT5ForConditionalGeneration', 'LongT5Model', 'LongT5PreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'FlaxLongT5ForConditionalGeneration', 'FlaxLongT5Model', 'FlaxLongT5PreTrainedModel', ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __UpperCAmelCase = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __lowercase : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : List[Any] ,A : int ,A : Optional[Any]=13 ,A : Dict=7 ,A : Dict=True ,A : Any=False ,A : Dict=99 ,A : int=32 ,A : Optional[int]=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : str=0.1 ,A : int=0.1 ,A : Optional[int]=20 ,A : Tuple=2 ,A : str=1 ,A : Optional[Any]=0 ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : int = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Union[str, Any] = pad_token_id UpperCAmelCase__ : List[str] = bos_token_id def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] ,self.vocab_size ).clip(3 ,self.vocab_size ) UpperCAmelCase__ : List[str] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) ,1 ) UpperCAmelCase__ : Any = np.concatenate([input_ids, eos_tensor] ,axis=1 ) UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : str = self.config_cls( vocab_size=self.vocab_size ,d_model=self.hidden_size ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,eos_token_ids=[2] ,bos_token_id=self.bos_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.pad_token_id ,**self.config_updates ,) UpperCAmelCase__ : Optional[Any] = prepare_pegasus_inputs_dict(A ,A ,A ) return config, inputs_dict def __lowercase ( self : Any ,A : Optional[int] ,A : str ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : Dict = model_class_name(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) ,dtype="""i4""" ) UpperCAmelCase__ : str = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Optional[int] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : int = model.decode( decoder_input_ids[:, -1:] ,A ,decoder_attention_mask=A ,past_key_values=outputs_cache.past_key_values ,decoder_position_ids=A ,) UpperCAmelCase__ : Dict = model.decode(A ,A ) UpperCAmelCase__ : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def __lowercase ( self : Optional[int] ,A : str ,A : Dict ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : str = model_class_name(A ) UpperCAmelCase__ : Any = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] ,axis=-1 ,) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : Dict = model.decode( decoder_input_ids[:, -1:] ,A ,past_key_values=outputs_cache.past_key_values ,decoder_attention_mask=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Union[str, Any] = model.decode(A ,A ,decoder_attention_mask=A ) UpperCAmelCase__ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , ): '''simple docstring''' if attention_mask is None: UpperCAmelCase__ : Union[str, Any] = np.not_equal(__UpperCamelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Tuple = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) snake_case_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = FlaxPegasusModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ) def __lowercase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(A ,A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(A ,A ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : List[Any] = self._prepare_for_class(A ,A ) UpperCAmelCase__ : int = model_class(A ) @jax.jit def encode_jitted(A : Optional[int] ,A : Union[str, Any]=None ,**A : Optional[Any] ): return model.encode(input_ids=A ,attention_mask=A ) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : int = encode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : Dict = encode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : Dict = model_class(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ,inputs_dict["""attention_mask"""] ) UpperCAmelCase__ : Dict = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(A : List[Any] ,A : Any ,A : List[Any] ): return model.decode( decoder_input_ids=A ,decoder_attention_mask=A ,encoder_outputs=A ,) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : Tuple = decode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : str = decode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""google/pegasus-large""" ,from_pt=A ) UpperCAmelCase__ : Any = np.ones((1, 1) ) UpperCAmelCase__ : Optional[Any] = model(A ) self.assertIsNotNone(A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Optional[Any] = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Union[str, Any] = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] UpperCAmelCase__ : str = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] UpperCAmelCase__ : str = tokenizer(A ,return_tensors="""np""" ,truncation=A ,max_length=512 ,padding=A ) UpperCAmelCase__ : Union[str, Any] = model.generate(**A ,num_beams=2 ).sequences UpperCAmelCase__ : int = tokenizer.batch_decode(A ,skip_special_tokens=A ) assert tgt_text == decoded
65
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Union[str, Any] ,A : Dict=7 ,A : Optional[int]=3 ,A : List[str]=18 ,A : Union[str, Any]=30 ,A : Tuple=400 ,A : Dict=True ,A : List[str]=None ,A : str=True ,A : Optional[Any]=False ,A : Optional[Any]=True ,A : List[str]=True ,A : Optional[int]=[0.5, 0.5, 0.5] ,A : List[str]=[0.5, 0.5, 0.5] ,): '''simple docstring''' UpperCAmelCase__ : str = parent UpperCAmelCase__ : List[str] = batch_size UpperCAmelCase__ : List[str] = num_channels UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : Optional[int] = max_resolution UpperCAmelCase__ : str = do_resize UpperCAmelCase__ : Tuple = size if size is not None else {"""height""": 18, """width""": 20} UpperCAmelCase__ : List[str] = do_thumbnail UpperCAmelCase__ : Optional[int] = do_align_axis UpperCAmelCase__ : Union[str, Any] = do_pad UpperCAmelCase__ : Tuple = do_normalize UpperCAmelCase__ : Optional[Any] = image_mean UpperCAmelCase__ : List[Any] = image_std def __lowercase ( self : Optional[int] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = DonutImageProcessor if is_vision_available() else None def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = DonutImageProcessingTester(self ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_thumbnail""" ) ) self.assertTrue(hasattr(A ,"""do_align_long_axis""" ) ) self.assertTrue(hasattr(A ,"""do_pad""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 18, """width""": 20} ) UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=(42, 84) ) self.assertEqual(image_processor.size ,{"""height""": 84, """width""": 42} ) def __lowercase ( self : Dict ): '''simple docstring''' pass @is_flaky() def __lowercase ( self : int ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase__ : int = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Tuple = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : List[str] ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase__ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Optional[int] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : Any ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : List[Any] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,)
65
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) UpperCAmelCase__ : Union[str, Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" from collections import namedtuple import requests from lxml import html # type: ignore __UpperCAmelCase = namedtuple('covid_data', 'cases deaths recovered') def lowerCAmelCase ( __UpperCamelCase = "https://www.worldometers.info/coronavirus/" ): '''simple docstring''' UpperCAmelCase__ : int = """//div[@class = \"maincounter-number\"]/span/text()""" return covid_data(*html.fromstring(requests.get(__UpperCamelCase ).content ).xpath(__UpperCamelCase ) ) __UpperCAmelCase = 'Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}' print(fmt.format(*covid_stats()))
65
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase ( __UpperCamelCase = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCAmelCase__ : Dict = F"{olid} is not a valid Open Library olid" raise ValueError(__UpperCamelCase ) return requests.get(F"https://openlibrary.org/{new_olid}.json" ).json() def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } UpperCAmelCase__ : Dict = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase__ : str = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] UpperCAmelCase__ : Dict = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Dict = """, """.join(__UpperCamelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __UpperCAmelCase = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: __UpperCAmelCase = summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('\n'.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
65
1
"""simple docstring""" import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if "cls_token" in name: UpperCAmelCase__ : Any = name.replace("""cls_token""" , """vit.embeddings.cls_token""" ) if "mask_token" in name: UpperCAmelCase__ : Optional[Any] = name.replace("""mask_token""" , """decoder.mask_token""" ) if "decoder_pos_embed" in name: UpperCAmelCase__ : int = name.replace("""decoder_pos_embed""" , """decoder.decoder_pos_embed""" ) if "pos_embed" in name and "decoder" not in name: UpperCAmelCase__ : List[str] = name.replace("""pos_embed""" , """vit.embeddings.position_embeddings""" ) if "patch_embed.proj" in name: UpperCAmelCase__ : List[str] = name.replace("""patch_embed.proj""" , """vit.embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: UpperCAmelCase__ : List[str] = name.replace("""patch_embed.norm""" , """vit.embeddings.norm""" ) if "decoder_blocks" in name: UpperCAmelCase__ : Tuple = name.replace("""decoder_blocks""" , """decoder.decoder_layers""" ) if "blocks" in name: UpperCAmelCase__ : Union[str, Any] = name.replace("""blocks""" , """vit.encoder.layer""" ) if "attn.proj" in name: UpperCAmelCase__ : Dict = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: UpperCAmelCase__ : List[Any] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: UpperCAmelCase__ : List[Any] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: UpperCAmelCase__ : Optional[Any] = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: UpperCAmelCase__ : Optional[int] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: UpperCAmelCase__ : List[str] = name.replace("""mlp.fc2""" , """output.dense""" ) if "decoder_embed" in name: UpperCAmelCase__ : Optional[Any] = name.replace("""decoder_embed""" , """decoder.decoder_embed""" ) if "decoder_norm" in name: UpperCAmelCase__ : List[str] = name.replace("""decoder_norm""" , """decoder.decoder_norm""" ) if "decoder_pred" in name: UpperCAmelCase__ : List[Any] = name.replace("""decoder_pred""" , """decoder.decoder_pred""" ) if "norm.weight" in name and "decoder" not in name: UpperCAmelCase__ : Any = name.replace("""norm.weight""" , """vit.layernorm.weight""" ) if "norm.bias" in name and "decoder" not in name: UpperCAmelCase__ : List[str] = name.replace("""norm.bias""" , """vit.layernorm.bias""" ) return name def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' for key in orig_state_dict.copy().keys(): UpperCAmelCase__ : Tuple = orig_state_dict.pop(__UpperCamelCase ) if "qkv" in key: UpperCAmelCase__ : Optional[int] = key.split(""".""" ) UpperCAmelCase__ : Union[str, Any] = int(key_split[1] ) if "decoder_blocks" in key: UpperCAmelCase__ : Optional[int] = config.decoder_hidden_size UpperCAmelCase__ : Optional[int] = """decoder.decoder_layers.""" if "weight" in key: UpperCAmelCase__ : Union[str, Any] = val[:dim, :] UpperCAmelCase__ : Any = val[dim : dim * 2, :] UpperCAmelCase__ : Optional[Any] = val[-dim:, :] elif "bias" in key: UpperCAmelCase__ : Optional[int] = val[:dim] UpperCAmelCase__ : Any = val[dim : dim * 2] UpperCAmelCase__ : List[Any] = val[-dim:] else: UpperCAmelCase__ : Union[str, Any] = config.hidden_size UpperCAmelCase__ : List[str] = """vit.encoder.layer.""" if "weight" in key: UpperCAmelCase__ : int = val[:dim, :] UpperCAmelCase__ : Dict = val[dim : dim * 2, :] UpperCAmelCase__ : Any = val[-dim:, :] elif "bias" in key: UpperCAmelCase__ : str = val[:dim] UpperCAmelCase__ : Tuple = val[dim : dim * 2] UpperCAmelCase__ : Dict = val[-dim:] else: UpperCAmelCase__ : Tuple = val return orig_state_dict def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ViTMAEConfig() if "large" in checkpoint_url: UpperCAmelCase__ : int = 1024 UpperCAmelCase__ : List[str] = 4096 UpperCAmelCase__ : List[Any] = 24 UpperCAmelCase__ : Any = 16 elif "huge" in checkpoint_url: UpperCAmelCase__ : List[str] = 14 UpperCAmelCase__ : int = 1280 UpperCAmelCase__ : int = 5120 UpperCAmelCase__ : int = 32 UpperCAmelCase__ : List[Any] = 16 UpperCAmelCase__ : str = ViTMAEForPreTraining(__UpperCamelCase ) UpperCAmelCase__ : Any = torch.hub.load_state_dict_from_url(__UpperCamelCase , map_location="""cpu""" )["""model"""] UpperCAmelCase__ : Dict = ViTMAEImageProcessor(size=config.image_size ) UpperCAmelCase__ : Optional[Any] = convert_state_dict(__UpperCamelCase , __UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) model.eval() UpperCAmelCase__ : List[Any] = """https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg""" UpperCAmelCase__ : int = Image.open(requests.get(__UpperCamelCase , stream=__UpperCamelCase ).raw ) UpperCAmelCase__ : Tuple = ViTMAEImageProcessor(size=config.image_size ) UpperCAmelCase__ : List[Any] = image_processor(images=__UpperCamelCase , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) UpperCAmelCase__ : Any = model(**__UpperCamelCase ) UpperCAmelCase__ : int = outputs.logits if "large" in checkpoint_url: UpperCAmelCase__ : Optional[int] = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: UpperCAmelCase__ : List[str] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: UpperCAmelCase__ : List[Any] = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __UpperCamelCase , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__UpperCamelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) __UpperCAmelCase = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
65
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=True , __UpperCamelCase="pt" ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {"""add_prefix_space""": True} if isinstance(__UpperCamelCase , __UpperCamelCase ) and not line.startswith(""" """ ) else {} UpperCAmelCase__ : List[str] = padding_side return tokenizer( [line] , max_length=__UpperCamelCase , padding="""max_length""" if pad_to_max_length else None , truncation=__UpperCamelCase , return_tensors=__UpperCamelCase , add_special_tokens=__UpperCamelCase , **__UpperCamelCase , ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , ): '''simple docstring''' UpperCAmelCase__ : str = input_ids.ne(__UpperCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ,A : List[Any] ,A : Union[str, Any] ,A : Any ,A : Optional[int] ,A : Union[str, Any]="train" ,A : Tuple=None ,A : Union[str, Any]=None ,A : Tuple=None ,A : int="" ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : Optional[Any] = Path(A ).joinpath(type_path + """.source""" ) UpperCAmelCase__ : List[str] = Path(A ).joinpath(type_path + """.target""" ) UpperCAmelCase__ : Dict = self.get_char_lens(self.src_file ) UpperCAmelCase__ : int = max_source_length UpperCAmelCase__ : List[str] = max_target_length assert min(self.src_lens ) > 0, f"found empty line in {self.src_file}" UpperCAmelCase__ : Dict = tokenizer UpperCAmelCase__ : str = prefix if n_obs is not None: UpperCAmelCase__ : int = self.src_lens[:n_obs] UpperCAmelCase__ : Any = src_lang UpperCAmelCase__ : Any = tgt_lang def __len__( self : Optional[Any] ): '''simple docstring''' return len(self.src_lens ) def __getitem__( self : Union[str, Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = index + 1 # linecache starts at 1 UpperCAmelCase__ : Tuple = self.prefix + linecache.getline(str(self.src_file ) ,A ).rstrip("""\n""" ) UpperCAmelCase__ : Dict = linecache.getline(str(self.tgt_file ) ,A ).rstrip("""\n""" ) assert source_line, f"empty source line for index {index}" assert tgt_line, f"empty tgt line for index {index}" # Need to add eos token manually for T5 if isinstance(self.tokenizer ,A ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right UpperCAmelCase__ : str = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,A ) else self.tokenizer ) UpperCAmelCase__ : Tuple = self.tokenizer.generator if isinstance(self.tokenizer ,A ) else self.tokenizer UpperCAmelCase__ : Tuple = encode_line(A ,A ,self.max_source_length ,"""right""" ) UpperCAmelCase__ : Dict = encode_line(A ,A ,self.max_target_length ,"""right""" ) UpperCAmelCase__ : Optional[Any] = source_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : List[str] = target_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : Union[str, Any] = source_inputs["""attention_mask"""].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def __lowercase ( A : int ): '''simple docstring''' return [len(A ) for x in Path(A ).open().readlines()] def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : int = torch.stack([x["""input_ids"""] for x in batch] ) UpperCAmelCase__ : Union[str, Any] = torch.stack([x["""attention_mask"""] for x in batch] ) UpperCAmelCase__ : Any = torch.stack([x["""decoder_input_ids"""] for x in batch] ) UpperCAmelCase__ : List[Any] = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : Any = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : str = trim_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = trim_batch(A ,A ,attention_mask=A ) UpperCAmelCase__ : List[str] = { """input_ids""": source_ids, """attention_mask""": source_mask, """decoder_input_ids""": y, } return batch __UpperCAmelCase = getLogger(__name__) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return list(itertools.chain.from_iterable(__UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = get_git_info() save_json(__UpperCamelCase , os.path.join(__UpperCamelCase , """git_log.json""" ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=4 , **__UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=__UpperCamelCase , **__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = git.Repo(search_parent_directories=__UpperCamelCase ) UpperCAmelCase__ : List[str] = { """repo_id""": str(__UpperCamelCase ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), """hostname""": str(socket.gethostname() ), } return repo_infos def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return list(map(__UpperCamelCase , __UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """wb""" ) as f: return pickle.dump(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' def remove_articles(__UpperCamelCase ): return re.sub(r"""\b(a|an|the)\b""" , """ """ , __UpperCamelCase ) def white_space_fix(__UpperCamelCase ): return " ".join(text.split() ) def remove_punc(__UpperCamelCase ): UpperCAmelCase__ : List[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCamelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCamelCase ) ) ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : Dict = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : int = Counter(__UpperCamelCase ) & Counter(__UpperCamelCase ) UpperCAmelCase__ : List[str] = sum(common.values() ) if num_same == 0: return 0 UpperCAmelCase__ : str = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Tuple = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return normalize_answer(__UpperCamelCase ) == normalize_answer(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' assert len(__UpperCamelCase ) == len(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 0 for hypo, pred in zip(__UpperCamelCase , __UpperCamelCase ): em += exact_match_score(__UpperCamelCase , __UpperCamelCase ) if len(__UpperCamelCase ) > 0: em /= len(__UpperCamelCase ) return {"em": em} def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return model_prefix.startswith("""rag""" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead UpperCAmelCase__ : str = """dropout_rate""" for p in extra_params: if getattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): if not hasattr(__UpperCamelCase , __UpperCamelCase ) and not hasattr(__UpperCamelCase , equivalent_param[p] ): logger.info("""config doesn't have a `{}` attribute""".format(__UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) continue UpperCAmelCase__ : Tuple = p if hasattr(__UpperCamelCase , __UpperCamelCase ) else equivalent_param[p] setattr(__UpperCamelCase , __UpperCamelCase , getattr(__UpperCamelCase , __UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) return hparams, config
65
1
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowercase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): @register_to_config def __init__( self : Tuple ,A : int ,A : int ,A : int ,A : float ,A : int ,A : int ,A : int ,A : int ,A : str ,A : bool = False ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : str = nn.Embedding(A ,A ) UpperCAmelCase__ : List[Any] = nn.Embedding(A ,A ) UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : List[str] = nn.Dropout(p=A ) UpperCAmelCase__ : List[Any] = TaConfig( vocab_size=A ,d_model=A ,num_heads=A ,d_kv=A ,d_ff=A ,dropout_rate=A ,feed_forward_proj=A ,is_decoder=A ,is_encoder_decoder=A ,) UpperCAmelCase__ : Dict = nn.ModuleList() for lyr_num in range(A ): UpperCAmelCase__ : int = TaBlock(A ) self.encoders.append(A ) UpperCAmelCase__ : Optional[Any] = TaLayerNorm(A ) UpperCAmelCase__ : int = nn.Dropout(p=A ) def __lowercase ( self : Union[str, Any] ,A : Union[str, Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.token_embedder(A ) UpperCAmelCase__ : List[Any] = encoder_input_tokens.shape[1] UpperCAmelCase__ : str = torch.arange(A ,device=encoder_input_tokens.device ) x += self.position_encoding(A ) UpperCAmelCase__ : Dict = self.dropout_pre(A ) # inverted the attention mask UpperCAmelCase__ : Any = encoder_input_tokens.size() UpperCAmelCase__ : Any = self.get_extended_attention_mask(A ,A ) for lyr in self.encoders: UpperCAmelCase__ : str = lyr(A ,A )[0] UpperCAmelCase__ : List[str] = self.layer_norm(A ) return self.dropout_post(A ), encoder_inputs_mask
65
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaControlnetPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return 32 @property def __lowercase ( self : int ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self : Any ): '''simple docstring''' return 100 @property def __lowercase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Tuple = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } UpperCAmelCase__ : int = UNetaDConditionModel(**A ) return model @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : str = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : str = self.dummy_unet UpperCAmelCase__ : List[Any] = self.dummy_movq UpperCAmelCase__ : List[Any] = DDIMScheduler( num_train_timesteps=1_000 ,beta_schedule="""linear""" ,beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,clip_sample=A ,set_alpha_to_one=A ,steps_offset=1 ,prediction_type="""epsilon""" ,thresholding=A ,) UpperCAmelCase__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __lowercase ( self : str ,A : Optional[Any] ,A : Any=0 ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( A ) # create hint UpperCAmelCase__ : int = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) if str(A ).startswith("""mps""" ): UpperCAmelCase__ : Optional[int] = torch.manual_seed(A ) else: UpperCAmelCase__ : Dict = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase__ : Dict = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = """cpu""" UpperCAmelCase__ : List[Any] = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = self.pipeline_class(**A ) UpperCAmelCase__ : Optional[int] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = pipe(**self.get_dummy_inputs(A ) ) UpperCAmelCase__ : Tuple = output.images UpperCAmelCase__ : Dict = pipe( **self.get_dummy_inputs(A ) ,return_dict=A ,)[0] UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1] UpperCAmelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Optional[int] = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : Union[str, Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) UpperCAmelCase__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) UpperCAmelCase__ : int = torch.from_numpy(np.array(A ) ).float() / 2_5_5.0 UpperCAmelCase__ : Union[str, Any] = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) UpperCAmelCase__ : List[str] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" ,torch_dtype=torch.floataa ) pipe_prior.to(A ) UpperCAmelCase__ : List[Any] = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" ,torch_dtype=torch.floataa ) UpperCAmelCase__ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[Any] = """A robot, 4k photo""" UpperCAmelCase__ : List[Any] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = pipe_prior( A ,generator=A ,num_inference_steps=5 ,negative_prompt="""""" ,).to_tuple() UpperCAmelCase__ : List[str] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ : int = pipeline( image_embeds=A ,negative_image_embeds=A ,hint=A ,generator=A ,num_inference_steps=100 ,output_type="""np""" ,) UpperCAmelCase__ : Any = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A ,A )
65
1
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece_bpe_char.model') @require_sentencepiece @require_tokenizers class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = SpeechTaTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Union[str, Any] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ : List[Any] = SpeechTaTokenizer(A ) UpperCAmelCase__ : int = AddedToken("""<mask>""" ,lstrip=A ,rstrip=A ) UpperCAmelCase__ : int = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """this is a test""" UpperCAmelCase__ : Optional[int] = """this is a test""" return input_text, output_text def __lowercase ( self : Optional[int] ,A : Dict ,A : Dict=False ,A : str=20 ,A : Dict=5 ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : List[str] = self.get_input_output_texts(A ) UpperCAmelCase__ : int = tokenizer.encode(A ,add_special_tokens=A ) UpperCAmelCase__ : Tuple = tokenizer.decode(A ,clean_up_tokenization_spaces=A ) return text, ids def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Tuple = """<pad>""" UpperCAmelCase__ : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[str] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""<s>""" ) self.assertEqual(vocab_keys[1] ,"""<pad>""" ) self.assertEqual(vocab_keys[-4] ,"""œ""" ) self.assertEqual(vocab_keys[-2] ,"""<mask>""" ) self.assertEqual(vocab_keys[-1] ,"""<ctc_blank>""" ) self.assertEqual(len(A ) ,81 ) def __lowercase ( self : Any ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,79 ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_tokenizers(do_lower_case=A ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): UpperCAmelCase__ : Optional[Any] = tokenizer.vocab_size UpperCAmelCase__ : Any = len(A ) self.assertNotEqual(A ,0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) UpperCAmelCase__ : List[Any] = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] UpperCAmelCase__ : Optional[Any] = tokenizer.add_tokens(A ) UpperCAmelCase__ : Dict = tokenizer.vocab_size UpperCAmelCase__ : Tuple = len(A ) self.assertNotEqual(A ,0 ) self.assertEqual(A ,A ) self.assertEqual(A ,len(A ) ) self.assertEqual(A ,all_size + len(A ) ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" ,add_special_tokens=A ) self.assertGreaterEqual(len(A ) ,4 ) self.assertGreater(tokens[0] ,tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] ,tokenizer.vocab_size - 1 ) UpperCAmelCase__ : List[Any] = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} UpperCAmelCase__ : Optional[Any] = tokenizer.add_special_tokens(A ) UpperCAmelCase__ : Any = tokenizer.vocab_size UpperCAmelCase__ : Optional[Any] = len(A ) self.assertNotEqual(A ,0 ) self.assertEqual(A ,A ) self.assertEqual(A ,len(A ) ) self.assertEqual(A ,all_size_a + len(A ) ) UpperCAmelCase__ : str = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" ,add_special_tokens=A ) self.assertGreaterEqual(len(A ) ,6 ) self.assertGreater(tokens[0] ,tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] ,tokens[1] ) self.assertGreater(tokens[-3] ,tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] ,tokens[-4] ) self.assertEqual(tokens[0] ,tokenizer.eos_token_id ) self.assertEqual(tokens[-3] ,tokenizer.pad_token_id ) def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_tokenizer() UpperCAmelCase__ : Tuple = tokenizer.tokenize("""This is a test""" ) # fmt: off self.assertListEqual(A ,[SPIECE_UNDERLINE, """T""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """a""", SPIECE_UNDERLINE, """t""", """e""", """s""", """t"""] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(A ) ,[4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] ,) UpperCAmelCase__ : Tuple = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( A ,[SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """92000""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) UpperCAmelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(A ) # fmt: off self.assertListEqual(A ,[4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on UpperCAmelCase__ : Optional[int] = tokenizer.convert_ids_to_tokens(A ) self.assertListEqual( A ,[SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """<unk>""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) @slow def __lowercase ( self : str ): '''simple docstring''' # Use custom sequence because this tokenizer does not handle numbers. UpperCAmelCase__ : Optional[Any] = [ """Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides """ """general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural """ """Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained """ """models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.""", """BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly """ """conditioning on both left and right context in all layers.""", """The quick brown fox jumps over the lazy dog.""", ] # fmt: off UpperCAmelCase__ : Optional[Any] = { """input_ids""": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""microsoft/speecht5_asr""" ,revision="""c5ef64c71905caeccde0e4462ef3f9077224c524""" ,sequences=A ,)
65
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = """vision-encoder-decoder""" snake_case_ = True def __init__( self : List[Any] ,**A : Union[str, Any] ): '''simple docstring''' super().__init__(**A ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) UpperCAmelCase__ : int = kwargs.pop("""encoder""" ) UpperCAmelCase__ : int = encoder_config.pop("""model_type""" ) UpperCAmelCase__ : str = kwargs.pop("""decoder""" ) UpperCAmelCase__ : Dict = decoder_config.pop("""model_type""" ) UpperCAmelCase__ : List[Any] = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Any = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Union[str, Any] = True @classmethod def __lowercase ( cls : List[Any] ,A : PretrainedConfig ,A : PretrainedConfig ,**A : Tuple ): '''simple docstring''' logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : List[Any] = True return cls(encoder=encoder_config.to_dict() ,decoder=decoder_config.to_dict() ,**A ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Dict = self.encoder.to_dict() UpperCAmelCase__ : Any = self.decoder.to_dict() UpperCAmelCase__ : Dict = self.__class__.model_type return output class __lowercase ( __lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' return 1e-4 @property def __lowercase ( self : List[Any] ): '''simple docstring''' return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : Optional[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : List[str] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def __lowercase ( self : Dict ,A : "PreTrainedTokenizerBase" ,A : int = -1 ,A : int = -1 ,A : bool = False ,A : Optional["TensorType"] = None ,): '''simple docstring''' import torch UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : List[Any] = super().generate_dummy_inputs( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) UpperCAmelCase__ , UpperCAmelCase__ : int = dummy_input["""input_ids"""].shape UpperCAmelCase__ : int = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCAmelCase__ : Tuple = dummy_input.pop("""input_ids""" ) UpperCAmelCase__ : Optional[int] = dummy_input.pop("""attention_mask""" ) UpperCAmelCase__ : Dict = torch.zeros(A ) return common_inputs class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : Any ,A : PretrainedConfig ): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(A ) def __lowercase ( self : Dict ,A : PretrainedConfig ,A : PretrainedConfig ,A : str = "default" ): '''simple docstring''' UpperCAmelCase__ : List[str] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(A ,A )
65
1
"""simple docstring""" import os import pytest from attr import dataclass __UpperCAmelCase = 'us-east-1' # defaults region @dataclass class __lowercase : snake_case_ = 42 snake_case_ = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" snake_case_ = { """task_name""": """mnli""", """per_device_train_batch_size""": 1_6, """per_device_eval_batch_size""": 1_6, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 5_0_0, """save_steps""": 5_5_0_0, } snake_case_ = {**hyperparameters, """max_steps""": 1_0_0_0} @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def __lowercase ( self : List[str] ): '''simple docstring''' return f"{self.framework}-transfromers-test" @property def __lowercase ( self : str ): '''simple docstring''' return f"./tests/sagemaker/scripts/{self.framework}" @property def __lowercase ( self : Tuple ): '''simple docstring''' if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="""class""" ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = SageMakerTestEnvironment(framework=request.cls.framework )
65
"""simple docstring""" import requests def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = {"""Content-Type""": """application/json"""} UpperCAmelCase__ : Optional[Any] = requests.post(__UpperCamelCase , json={"""text""": message_body} , headers=__UpperCamelCase ) if response.status_code != 200: UpperCAmelCase__ : Any = ( """Request to slack returned an error """ F"{response.status_code}, the response is:\n{response.text}" ) raise ValueError(__UpperCamelCase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
65
1
"""simple docstring""" from math import isqrt def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return all(number % divisor != 0 for divisor in range(2 , isqrt(__UpperCamelCase ) + 1 ) ) def lowerCAmelCase ( __UpperCamelCase = 10**6 ): '''simple docstring''' UpperCAmelCase__ : str = 0 UpperCAmelCase__ : Dict = 1 UpperCAmelCase__ : Optional[Any] = 7 while prime_candidate < max_prime: primes_count += is_prime(__UpperCamelCase ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"{solution() = }")
65
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = CTRLTokenizer snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ : List[Any] = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase__ : Optional[int] = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : List[Any] = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase__ : int = {"""unk_token""": """<unk>"""} UpperCAmelCase__ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(A ) ) def __lowercase ( self : int ,**A : Dict ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """adapt react readapt apt""" UpperCAmelCase__ : Any = """adapt react readapt apt""" return input_text, output_text def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = CTRLTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) UpperCAmelCase__ : Tuple = """adapt react readapt apt""" UpperCAmelCase__ : Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase__ : Dict = tokenizer.tokenize(A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Any = tokens + [tokenizer.unk_token] UpperCAmelCase__ : Dict = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,A )
65
1
"""simple docstring""" import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } __UpperCAmelCase = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' for attribute in key.split(""".""" ): UpperCAmelCase__ : List[Any] = getattr(__UpperCamelCase , __UpperCamelCase ) if weight_type is not None: UpperCAmelCase__ : Optional[int] = getattr(__UpperCamelCase , __UpperCamelCase ).shape else: UpperCAmelCase__ : int = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": UpperCAmelCase__ : str = value elif weight_type == "weight_g": UpperCAmelCase__ : Dict = value elif weight_type == "weight_v": UpperCAmelCase__ : List[str] = value elif weight_type == "bias": UpperCAmelCase__ : Optional[Any] = value else: UpperCAmelCase__ : Optional[Any] = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = [] UpperCAmelCase__ : int = fairseq_model.state_dict() UpperCAmelCase__ : Any = hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase__ : Optional[Any] = False if "conv_layers" in name: load_conv_layer( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , hf_model.config.feat_extract_norm == """group""" , ) UpperCAmelCase__ : Union[str, Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: UpperCAmelCase__ : List[str] = True if "*" in mapped_key: UpperCAmelCase__ : Any = name.split(__UpperCamelCase )[0].split(""".""" )[-2] UpperCAmelCase__ : Optional[int] = mapped_key.replace("""*""" , __UpperCamelCase ) if "weight_g" in name: UpperCAmelCase__ : str = """weight_g""" elif "weight_v" in name: UpperCAmelCase__ : Optional[Any] = """weight_v""" elif "bias" in name and "relative_attention_bias" not in name: UpperCAmelCase__ : Dict = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase__ : Tuple = """weight""" else: UpperCAmelCase__ : List[Any] = None set_recursively(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) continue if not is_used: unused_weights.append(__UpperCamelCase ) logger.warning(F"Unused weights: {unused_weights}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = full_name.split("""conv_layers.""" )[-1] UpperCAmelCase__ : Tuple = name.split(""".""" ) UpperCAmelCase__ : Optional[Any] = int(items[0] ) UpperCAmelCase__ : int = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) UpperCAmelCase__ : int = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) UpperCAmelCase__ : Tuple = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) UpperCAmelCase__ : Optional[int] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) UpperCAmelCase__ : Tuple = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(__UpperCamelCase ) @torch.no_grad() def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None ): '''simple docstring''' UpperCAmelCase__ : str = torch.load(__UpperCamelCase ) UpperCAmelCase__ : List[str] = WavLMConfigOrig(checkpoint["""cfg"""] ) UpperCAmelCase__ : Optional[Any] = WavLMOrig(__UpperCamelCase ) model.load_state_dict(checkpoint["""model"""] ) model.eval() if config_path is not None: UpperCAmelCase__ : Tuple = WavLMConfig.from_pretrained(__UpperCamelCase ) else: UpperCAmelCase__ : List[Any] = WavLMConfig() UpperCAmelCase__ : Optional[int] = WavLMModel(__UpperCamelCase ) recursively_load_weights(__UpperCamelCase , __UpperCamelCase ) hf_wavlm.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') __UpperCAmelCase = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
65
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCAmelCase = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfig', 'BridgeTowerTextConfig', 'BridgeTowerVisionConfig', ], 'processing_bridgetower': ['BridgeTowerProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['BridgeTowerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST', 'BridgeTowerForContrastiveLearning', 'BridgeTowerForImageAndTextRetrieval', 'BridgeTowerForMaskedLM', 'BridgeTowerModel', 'BridgeTowerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
65
1
"""simple docstring""" from __future__ import annotations __UpperCAmelCase = 1.6021E-19 # units = C def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): '''simple docstring''' if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif conductivity < 0: raise ValueError("""Conductivity cannot be negative""" ) elif electron_conc < 0: raise ValueError("""Electron concentration cannot be negative""" ) elif mobility < 0: raise ValueError("""mobility cannot be negative""" ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
65
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """is_longer"""] def __init__( self : str ,A : Union[str, Any]=64 ,A : Tuple=48_000 ,A : Dict=480 ,A : List[str]=10 ,A : str=1_024 ,A : Any=0.0 ,A : Optional[int]=False ,A : float = 0 ,A : float = 14_000 ,A : int = None ,A : str = "fusion" ,A : str = "repeatpad" ,**A : List[Any] ,): '''simple docstring''' super().__init__( feature_size=A ,sampling_rate=A ,padding_value=A ,return_attention_mask=A ,**A ,) UpperCAmelCase__ : List[Any] = top_db UpperCAmelCase__ : Union[str, Any] = truncation UpperCAmelCase__ : Optional[int] = padding UpperCAmelCase__ : List[Any] = fft_window_size UpperCAmelCase__ : Optional[Any] = (fft_window_size >> 1) + 1 UpperCAmelCase__ : Any = hop_length UpperCAmelCase__ : List[str] = max_length_s UpperCAmelCase__ : List[Any] = max_length_s * sampling_rate UpperCAmelCase__ : List[Any] = sampling_rate UpperCAmelCase__ : Optional[int] = frequency_min UpperCAmelCase__ : Tuple = frequency_max UpperCAmelCase__ : List[str] = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm=A ,mel_scale="""htk""" ,) UpperCAmelCase__ : str = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm="""slaney""" ,mel_scale="""slaney""" ,) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def __lowercase ( self : List[str] ,A : np.array ,A : Optional[np.array] = None ): '''simple docstring''' UpperCAmelCase__ : Dict = spectrogram( A ,window_function(self.fft_window_size ,"""hann""" ) ,frame_length=self.fft_window_size ,hop_length=self.hop_length ,power=2.0 ,mel_filters=A ,log_mel="""dB""" ,) return log_mel_spectrogram.T def __lowercase ( self : Optional[Any] ,A : Union[str, Any] ,A : int ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = np.array_split(list(range(0 ,total_frames - chunk_frames + 1 ) ) ,3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : List[str] = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : int = [0] # randomly choose index for each part UpperCAmelCase__ : Tuple = np.random.choice(ranges[0] ) UpperCAmelCase__ : Tuple = np.random.choice(ranges[1] ) UpperCAmelCase__ : str = np.random.choice(ranges[2] ) UpperCAmelCase__ : List[str] = mel[idx_front : idx_front + chunk_frames, :] UpperCAmelCase__ : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] UpperCAmelCase__ : Dict = mel[idx_back : idx_back + chunk_frames, :] UpperCAmelCase__ : Optional[Any] = torch.tensor(mel[None, None, :] ) UpperCAmelCase__ : int = torch.nn.functional.interpolate( A ,size=[chunk_frames, 64] ,mode="""bilinear""" ,align_corners=A ) UpperCAmelCase__ : Dict = mel_shrink[0][0].numpy() UpperCAmelCase__ : Dict = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] ,axis=0 ) return mel_fusion def __lowercase ( self : Any ,A : np.array ,A : Optional[int] ,A : Any ,A : Tuple ): '''simple docstring''' if waveform.shape[0] > max_length: if truncation == "rand_trunc": UpperCAmelCase__ : int = True # random crop to max_length (for compatibility) -> this should be handled by self.pad UpperCAmelCase__ : str = len(A ) - max_length UpperCAmelCase__ : Optional[Any] = np.random.randint(0 ,overflow + 1 ) UpperCAmelCase__ : Optional[int] = waveform[idx : idx + max_length] UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] elif truncation == "fusion": UpperCAmelCase__ : Tuple = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : Optional[int] = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed UpperCAmelCase__ : int = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. UpperCAmelCase__ : List[Any] = np.stack([mel, mel, mel, mel] ,axis=0 ) UpperCAmelCase__ : Any = False else: UpperCAmelCase__ : Union[str, Any] = self._random_mel_fusion(A ,A ,A ) UpperCAmelCase__ : List[str] = True else: raise NotImplementedError(f"data_truncating {truncation} not implemented" ) else: UpperCAmelCase__ : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": UpperCAmelCase__ : str = int(max_length / len(A ) ) UpperCAmelCase__ : int = np.stack(np.tile(A ,n_repeat + 1 ) )[:max_length] if padding == "repeatpad": UpperCAmelCase__ : List[Any] = int(max_length / len(A ) ) UpperCAmelCase__ : str = np.stack(np.tile(A ,A ) ) UpperCAmelCase__ : Optional[Any] = np.pad(A ,(0, max_length - waveform.shape[0]) ,mode="""constant""" ,constant_values=0 ) if truncation == "fusion": UpperCAmelCase__ : int = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] ,axis=0 ) else: UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self : str ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : str = None ,A : Optional[str] = None ,A : Optional[int] = None ,A : Optional[int] = None ,A : Optional[Union[str, TensorType]] = None ,**A : List[str] ,): '''simple docstring''' UpperCAmelCase__ : Optional[int] = truncation if truncation is not None else self.truncation UpperCAmelCase__ : Dict = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" f" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" f" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[int] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : List[str] = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Any = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [np.asarray(A )] # convert to mel spectrogram, truncate and pad if needed. UpperCAmelCase__ : Tuple = [ self._get_input_mel(A ,max_length if max_length else self.nb_max_samples ,A ,A ) for waveform in raw_speech ] UpperCAmelCase__ : Optional[int] = [] UpperCAmelCase__ : Tuple = [] for mel, longer in padded_inputs: input_mel.append(A ) is_longer.append(A ) if truncation == "fusion" and sum(A ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer UpperCAmelCase__ : List[str] = np.random.randint(0 ,len(A ) ) UpperCAmelCase__ : int = True if isinstance(input_mel[0] ,A ): UpperCAmelCase__ : Tuple = [np.asarray(A ,dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool UpperCAmelCase__ : List[str] = [[longer] for longer in is_longer] UpperCAmelCase__ : List[Any] = {"""input_features""": input_mel, """is_longer""": is_longer} UpperCAmelCase__ : str = BatchFeature(A ) if return_tensors is not None: UpperCAmelCase__ : int = input_features.convert_to_tensors(A ) return input_features
65
1
"""simple docstring""" import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if isinstance(__UpperCamelCase , torch.Tensor ): return image elif isinstance(__UpperCamelCase , PIL.Image.Image ): UpperCAmelCase__ : List[Any] = [image] if isinstance(image[0] , PIL.Image.Image ): UpperCAmelCase__ : Optional[int] = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] UpperCAmelCase__ : Union[str, Any] = np.concatenate(__UpperCamelCase , axis=0 ) UpperCAmelCase__ : str = np.array(__UpperCamelCase ).astype(np.floataa ) / 255.0 UpperCAmelCase__ : List[str] = image.transpose(0 , 3 , 1 , 2 ) UpperCAmelCase__ : Optional[int] = 2.0 * image - 1.0 UpperCAmelCase__ : List[Any] = torch.from_numpy(__UpperCamelCase ) elif isinstance(image[0] , torch.Tensor ): UpperCAmelCase__ : List[Any] = torch.cat(__UpperCamelCase , dim=0 ) return image def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=0.9995 ): '''simple docstring''' if not isinstance(__UpperCamelCase , np.ndarray ): UpperCAmelCase__ : Any = True UpperCAmelCase__ : Optional[int] = va.device UpperCAmelCase__ : Tuple = va.cpu().numpy() UpperCAmelCase__ : Optional[int] = va.cpu().numpy() UpperCAmelCase__ : Dict = np.sum(va * va / (np.linalg.norm(__UpperCamelCase ) * np.linalg.norm(__UpperCamelCase )) ) if np.abs(__UpperCamelCase ) > DOT_THRESHOLD: UpperCAmelCase__ : Tuple = (1 - t) * va + t * va else: UpperCAmelCase__ : str = np.arccos(__UpperCamelCase ) UpperCAmelCase__ : int = np.sin(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = theta_a * t UpperCAmelCase__ : int = np.sin(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = np.sin(theta_a - theta_t ) / sin_theta_a UpperCAmelCase__ : Tuple = sin_theta_t / sin_theta_a UpperCAmelCase__ : List[Any] = sa * va + sa * va if inputs_are_torch: UpperCAmelCase__ : Dict = torch.from_numpy(__UpperCamelCase ).to(__UpperCamelCase ) return va def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : str = F.normalize(__UpperCamelCase , dim=-1 ) UpperCAmelCase__ : Union[str, Any] = F.normalize(__UpperCamelCase , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' for param in model.parameters(): UpperCAmelCase__ : Any = value class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ,A : AutoencoderKL ,A : CLIPTextModel ,A : CLIPModel ,A : CLIPTokenizer ,A : UNetaDConditionModel ,A : Union[PNDMScheduler, LMSDiscreteScheduler, DDIMScheduler, DPMSolverMultistepScheduler] ,A : CLIPFeatureExtractor ,A : Optional[Any]=None ,A : Union[str, Any]=None ,A : str=None ,): '''simple docstring''' super().__init__() self.register_modules( vae=A ,text_encoder=A ,clip_model=A ,tokenizer=A ,unet=A ,scheduler=A ,feature_extractor=A ,coca_model=A ,coca_tokenizer=A ,coca_transform=A ,) UpperCAmelCase__ : List[Any] = ( feature_extractor.size if isinstance(feature_extractor.size ,A ) else feature_extractor.size["""shortest_edge"""] ) UpperCAmelCase__ : str = transforms.Normalize(mean=feature_extractor.image_mean ,std=feature_extractor.image_std ) set_requires_grad(self.text_encoder ,A ) set_requires_grad(self.clip_model ,A ) def __lowercase ( self : Optional[int] ,A : Optional[Union[str, int]] = "auto" ): '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory UpperCAmelCase__ : int = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(A ) def __lowercase ( self : int ): '''simple docstring''' self.enable_attention_slicing(A ) def __lowercase ( self : List[str] ): '''simple docstring''' set_requires_grad(self.vae ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' set_requires_grad(self.vae ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' set_requires_grad(self.unet ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' set_requires_grad(self.unet ,A ) def __lowercase ( self : Dict ,A : str ,A : List[Any] ,A : int ): '''simple docstring''' # get the original timestep using init_timestep UpperCAmelCase__ : Any = min(int(num_inference_steps * strength ) ,A ) UpperCAmelCase__ : List[Any] = max(num_inference_steps - init_timestep ,0 ) UpperCAmelCase__ : Any = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def __lowercase ( self : str ,A : Optional[int] ,A : Dict ,A : int ,A : Optional[int] ,A : Optional[Any] ,A : int=None ): '''simple docstring''' if not isinstance(A ,torch.Tensor ): raise ValueError(f"`image` has to be of type `torch.Tensor` but is {type(A )}" ) UpperCAmelCase__ : int = image.to(device=A ,dtype=A ) if isinstance(A ,A ): UpperCAmelCase__ : List[Any] = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(A ) ] UpperCAmelCase__ : Union[str, Any] = torch.cat(A ,dim=0 ) else: UpperCAmelCase__ : List[Any] = self.vae.encode(A ).latent_dist.sample(A ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor UpperCAmelCase__ : Any = 0.1_8_2_1_5 * init_latents UpperCAmelCase__ : Tuple = init_latents.repeat_interleave(A ,dim=0 ) UpperCAmelCase__ : Any = randn_tensor(init_latents.shape ,generator=A ,device=A ,dtype=A ) # get latents UpperCAmelCase__ : Optional[Any] = self.scheduler.add_noise(A ,A ,A ) UpperCAmelCase__ : Union[str, Any] = init_latents return latents def __lowercase ( self : List[Any] ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.coca_transform(A ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): UpperCAmelCase__ : Optional[Any] = self.coca_model.generate(transformed_image.to(device=self.device ,dtype=self.coca_model.dtype ) ) UpperCAmelCase__ : str = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("""<end_of_text>""" )[0].replace("""<start_of_text>""" ,"""""" ).rstrip(""" .,""" ) def __lowercase ( self : str ,A : List[str] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.feature_extractor.preprocess(A ) UpperCAmelCase__ : List[Any] = torch.from_numpy(clip_image_input["""pixel_values"""][0] ).unsqueeze(0 ).to(self.device ).half() UpperCAmelCase__ : Optional[Any] = self.clip_model.get_image_features(A ) UpperCAmelCase__ : Dict = image_embeddings_clip / image_embeddings_clip.norm(p=2 ,dim=-1 ,keepdim=A ) UpperCAmelCase__ : Tuple = image_embeddings_clip.repeat_interleave(A ,dim=0 ) return image_embeddings_clip @torch.enable_grad() def __lowercase ( self : Any ,A : List[Any] ,A : List[Any] ,A : int ,A : int ,A : int ,A : List[str] ,A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : Tuple = latents.detach().requires_grad_() UpperCAmelCase__ : Tuple = self.scheduler.scale_model_input(A ,A ) # predict the noise residual UpperCAmelCase__ : List[Any] = self.unet(A ,A ,encoder_hidden_states=A ).sample if isinstance(self.scheduler ,(PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): UpperCAmelCase__ : str = self.scheduler.alphas_cumprod[timestep] UpperCAmelCase__ : Any = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCAmelCase__ : Dict = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 UpperCAmelCase__ : int = torch.sqrt(A ) UpperCAmelCase__ : List[Any] = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler ,A ): UpperCAmelCase__ : List[Any] = self.scheduler.sigmas[index] UpperCAmelCase__ : Any = latents - sigma * noise_pred else: raise ValueError(f"scheduler type {type(self.scheduler )} not supported" ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor UpperCAmelCase__ : List[Any] = 1 / 0.1_8_2_1_5 * sample UpperCAmelCase__ : Union[str, Any] = self.vae.decode(A ).sample UpperCAmelCase__ : Optional[int] = (image / 2 + 0.5).clamp(0 ,1 ) UpperCAmelCase__ : Tuple = transforms.Resize(self.feature_extractor_size )(A ) UpperCAmelCase__ : List[Any] = self.normalize(A ).to(latents.dtype ) UpperCAmelCase__ : Union[str, Any] = self.clip_model.get_image_features(A ) UpperCAmelCase__ : Optional[int] = image_embeddings_clip / image_embeddings_clip.norm(p=2 ,dim=-1 ,keepdim=A ) UpperCAmelCase__ : Union[str, Any] = spherical_dist_loss(A ,A ).mean() * clip_guidance_scale UpperCAmelCase__ : List[Any] = -torch.autograd.grad(A ,A )[0] if isinstance(self.scheduler ,A ): UpperCAmelCase__ : List[str] = latents.detach() + grads * (sigma**2) UpperCAmelCase__ : Optional[Any] = noise_pred_original else: UpperCAmelCase__ : Tuple = noise_pred_original - torch.sqrt(A ) * grads return noise_pred, latents @torch.no_grad() def __call__( self : Dict ,A : Union[torch.FloatTensor, PIL.Image.Image] ,A : Union[torch.FloatTensor, PIL.Image.Image] ,A : Optional[str] = None ,A : Optional[str] = None ,A : Optional[int] = 512 ,A : Optional[int] = 512 ,A : float = 0.6 ,A : Optional[int] = 50 ,A : Optional[float] = 7.5 ,A : Optional[int] = 1 ,A : float = 0.0 ,A : Optional[float] = 100 ,A : Optional[torch.Generator] = None ,A : Optional[str] = "pil" ,A : bool = True ,A : float = 0.8 ,A : float = 0.1 ,A : float = 0.1 ,): '''simple docstring''' if isinstance(A ,A ) and len(A ) != batch_size: raise ValueError(f"You have passed {batch_size} batch_size, but only {len(A )} generators." ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f"`height` and `width` have to be divisible by 8 but are {height} and {width}." ) if isinstance(A ,torch.Generator ) and batch_size > 1: UpperCAmelCase__ : int = [generator] + [None] * (batch_size - 1) UpperCAmelCase__ : Union[str, Any] = [ ("""model""", self.coca_model is None), ("""tokenizer""", self.coca_tokenizer is None), ("""transform""", self.coca_transform is None), ] UpperCAmelCase__ : str = [x[0] for x in coca_is_none if x[1]] UpperCAmelCase__ : Optional[Any] = """, """.join(A ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(A ): raise ValueError( f"Content prompt is None and CoCa [{coca_is_none_str}] is None." f"Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline." ) UpperCAmelCase__ : Union[str, Any] = self.get_image_description(A ) if style_prompt is None: if len(A ): raise ValueError( f"Style prompt is None and CoCa [{coca_is_none_str}] is None." f" Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline." ) UpperCAmelCase__ : Optional[Any] = self.get_image_description(A ) # get prompt text embeddings for content and style UpperCAmelCase__ : Any = self.tokenizer( A ,padding="""max_length""" ,max_length=self.tokenizer.model_max_length ,truncation=A ,return_tensors="""pt""" ,) UpperCAmelCase__ : List[Any] = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] UpperCAmelCase__ : List[str] = self.tokenizer( A ,padding="""max_length""" ,max_length=self.tokenizer.model_max_length ,truncation=A ,return_tensors="""pt""" ,) UpperCAmelCase__ : List[str] = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] UpperCAmelCase__ : Tuple = slerp(A ,A ,A ) # duplicate text embeddings for each generation per prompt UpperCAmelCase__ : Any = text_embeddings.repeat_interleave(A ,dim=0 ) # set timesteps UpperCAmelCase__ : List[Any] = """offset""" in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) UpperCAmelCase__ : Any = {} if accepts_offset: UpperCAmelCase__ : List[Any] = 1 self.scheduler.set_timesteps(A ,**A ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.get_timesteps(A ,A ,self.device ) UpperCAmelCase__ : List[str] = timesteps[:1].repeat(A ) # Preprocess image UpperCAmelCase__ : Tuple = preprocess(A ,A ,A ) UpperCAmelCase__ : str = self.prepare_latents( A ,A ,A ,text_embeddings.dtype ,self.device ,A ) UpperCAmelCase__ : Tuple = preprocess(A ,A ,A ) UpperCAmelCase__ : Dict = self.prepare_latents( A ,A ,A ,text_embeddings.dtype ,self.device ,A ) UpperCAmelCase__ : int = slerp(A ,A ,A ) if clip_guidance_scale > 0: UpperCAmelCase__ : List[Any] = self.get_clip_image_embeddings(A ,A ) UpperCAmelCase__ : Any = self.get_clip_image_embeddings(A ,A ) UpperCAmelCase__ : Optional[Any] = slerp( A ,A ,A ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. UpperCAmelCase__ : Dict = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: UpperCAmelCase__ : Dict = content_text_input.input_ids.shape[-1] UpperCAmelCase__ : List[Any] = self.tokenizer([""""""] ,padding="""max_length""" ,max_length=A ,return_tensors="""pt""" ) UpperCAmelCase__ : Union[str, Any] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt UpperCAmelCase__ : Optional[int] = uncond_embeddings.repeat_interleave(A ,dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase__ : Optional[int] = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. UpperCAmelCase__ : Dict = (batch_size, self.unet.config.in_channels, height // 8, width // 8) UpperCAmelCase__ : List[Any] = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps UpperCAmelCase__ : Union[str, Any] = torch.randn(A ,generator=A ,device="""cpu""" ,dtype=A ).to( self.device ) else: UpperCAmelCase__ : Optional[int] = torch.randn(A ,generator=A ,device=self.device ,dtype=A ) else: if latents.shape != latents_shape: raise ValueError(f"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) UpperCAmelCase__ : List[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler UpperCAmelCase__ : Optional[int] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] UpperCAmelCase__ : Union[str, Any] = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCAmelCase__ : Optional[int] = {} if accepts_eta: UpperCAmelCase__ : Union[str, Any] = eta # check if the scheduler accepts generator UpperCAmelCase__ : str = """generator""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: UpperCAmelCase__ : Optional[int] = generator with self.progress_bar(total=A ): for i, t in enumerate(A ): # expand the latents if we are doing classifier free guidance UpperCAmelCase__ : List[str] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase__ : int = self.scheduler.scale_model_input(A ,A ) # predict the noise residual UpperCAmelCase__ : Tuple = self.unet(A ,A ,encoder_hidden_states=A ).sample # perform classifier free guidance if do_classifier_free_guidance: UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = noise_pred.chunk(2 ) UpperCAmelCase__ : List[str] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: UpperCAmelCase__ : Optional[int] = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.cond_fn( A ,A ,A ,A ,A ,A ,A ,) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase__ : List[Any] = self.scheduler.step(A ,A ,A ,**A ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor UpperCAmelCase__ : Any = 1 / 0.1_8_2_1_5 * latents UpperCAmelCase__ : int = self.vae.decode(A ).sample UpperCAmelCase__ : Any = (image / 2 + 0.5).clamp(0 ,1 ) UpperCAmelCase__ : Optional[Any] = image.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": UpperCAmelCase__ : int = self.numpy_to_pil(A ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=A ,nsfw_content_detected=A )
65
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Union[str, Any] ,A : Dict=7 ,A : Optional[int]=3 ,A : List[str]=18 ,A : Union[str, Any]=30 ,A : Tuple=400 ,A : Dict=True ,A : List[str]=None ,A : str=True ,A : Optional[Any]=False ,A : Optional[Any]=True ,A : List[str]=True ,A : Optional[int]=[0.5, 0.5, 0.5] ,A : List[str]=[0.5, 0.5, 0.5] ,): '''simple docstring''' UpperCAmelCase__ : str = parent UpperCAmelCase__ : List[str] = batch_size UpperCAmelCase__ : List[str] = num_channels UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : Optional[int] = max_resolution UpperCAmelCase__ : str = do_resize UpperCAmelCase__ : Tuple = size if size is not None else {"""height""": 18, """width""": 20} UpperCAmelCase__ : List[str] = do_thumbnail UpperCAmelCase__ : Optional[int] = do_align_axis UpperCAmelCase__ : Union[str, Any] = do_pad UpperCAmelCase__ : Tuple = do_normalize UpperCAmelCase__ : Optional[Any] = image_mean UpperCAmelCase__ : List[Any] = image_std def __lowercase ( self : Optional[int] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = DonutImageProcessor if is_vision_available() else None def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = DonutImageProcessingTester(self ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_thumbnail""" ) ) self.assertTrue(hasattr(A ,"""do_align_long_axis""" ) ) self.assertTrue(hasattr(A ,"""do_pad""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 18, """width""": 20} ) UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=(42, 84) ) self.assertEqual(image_processor.size ,{"""height""": 84, """width""": 42} ) def __lowercase ( self : Dict ): '''simple docstring''' pass @is_flaky() def __lowercase ( self : int ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase__ : int = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Tuple = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : List[str] ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase__ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Optional[int] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : Any ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : List[Any] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,)
65
1
"""simple docstring""" import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingface_hub.utils import ( EntryNotFoundError, RepositoryNotFoundError, RevisionNotFoundError, is_jinja_available, ) from packaging import version from requests import HTTPError from .. import __version__ from .constants import ( DEPRECATED_REVISION_ARGS, DIFFUSERS_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, SAFETENSORS_WEIGHTS_NAME, WEIGHTS_NAME, ) from .import_utils import ( ENV_VARS_TRUE_VALUES, _flax_version, _jax_version, _onnxruntime_version, _torch_version, is_flax_available, is_onnx_available, is_torch_available, ) from .logging import get_logger __UpperCAmelCase = get_logger(__name__) __UpperCAmelCase = Path(__file__).parent / 'model_card_template.md' __UpperCAmelCase = uuida().hex __UpperCAmelCase = os.getenv('HF_HUB_OFFLINE', '').upper() in ENV_VARS_TRUE_VALUES __UpperCAmelCase = os.getenv('DISABLE_TELEMETRY', '').upper() in ENV_VARS_TRUE_VALUES __UpperCAmelCase = HUGGINGFACE_CO_RESOLVE_ENDPOINT + '/api/telemetry/' def lowerCAmelCase ( __UpperCamelCase = None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = F"diffusers/{__version__}; python/{sys.version.split()[0]}; session_id/{SESSION_ID}" if DISABLE_TELEMETRY or HF_HUB_OFFLINE: return ua + "; telemetry/off" if is_torch_available(): ua += F"; torch/{_torch_version}" if is_flax_available(): ua += F"; jax/{_jax_version}" ua += F"; flax/{_flax_version}" if is_onnx_available(): ua += F"; onnxruntime/{_onnxruntime_version}" # CI will set this value to True if os.environ.get("""DIFFUSERS_IS_CI""" , """""" ).upper() in ENV_VARS_TRUE_VALUES: ua += "; is_ci/true" if isinstance(__UpperCamelCase , __UpperCamelCase ): ua += "; " + "; ".join(F"{k}/{v}" for k, v in user_agent.items() ) elif isinstance(__UpperCamelCase , __UpperCamelCase ): ua += "; " + user_agent return ua def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): '''simple docstring''' if token is None: UpperCAmelCase__ : List[str] = HfFolder.get_token() if organization is None: UpperCAmelCase__ : Optional[Any] = whoami(__UpperCamelCase )["""name"""] return F"{username}/{model_id}" else: return F"{organization}/{model_id}" def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if not is_jinja_available(): raise ValueError( """Modelcard rendering is based on Jinja templates.""" """ Please make sure to have `jinja` installed before using `create_model_card`.""" """ To install it, please run `pip install Jinja2`.""" ) if hasattr(__UpperCamelCase , """local_rank""" ) and args.local_rank not in [-1, 0]: return UpperCAmelCase__ : str = args.hub_token if hasattr(__UpperCamelCase , """hub_token""" ) else None UpperCAmelCase__ : int = get_full_repo_name(__UpperCamelCase , token=__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = ModelCard.from_template( card_data=ModelCardData( # Card metadata object that will be converted to YAML block language="""en""" , license="""apache-2.0""" , library_name="""diffusers""" , tags=[] , datasets=args.dataset_name , metrics=[] , ) , template_path=__UpperCamelCase , model_name=__UpperCamelCase , repo_name=__UpperCamelCase , dataset_name=args.dataset_name if hasattr(__UpperCamelCase , """dataset_name""" ) else None , learning_rate=args.learning_rate , train_batch_size=args.train_batch_size , eval_batch_size=args.eval_batch_size , gradient_accumulation_steps=( args.gradient_accumulation_steps if hasattr(__UpperCamelCase , """gradient_accumulation_steps""" ) else None ) , adam_betaa=args.adam_betaa if hasattr(__UpperCamelCase , """adam_beta1""" ) else None , adam_betaa=args.adam_betaa if hasattr(__UpperCamelCase , """adam_beta2""" ) else None , adam_weight_decay=args.adam_weight_decay if hasattr(__UpperCamelCase , """adam_weight_decay""" ) else None , adam_epsilon=args.adam_epsilon if hasattr(__UpperCamelCase , """adam_epsilon""" ) else None , lr_scheduler=args.lr_scheduler if hasattr(__UpperCamelCase , """lr_scheduler""" ) else None , lr_warmup_steps=args.lr_warmup_steps if hasattr(__UpperCamelCase , """lr_warmup_steps""" ) else None , ema_inv_gamma=args.ema_inv_gamma if hasattr(__UpperCamelCase , """ema_inv_gamma""" ) else None , ema_power=args.ema_power if hasattr(__UpperCamelCase , """ema_power""" ) else None , ema_max_decay=args.ema_max_decay if hasattr(__UpperCamelCase , """ema_max_decay""" ) else None , mixed_precision=args.mixed_precision , ) UpperCAmelCase__ : List[str] = os.path.join(args.output_dir , """README.md""" ) model_card.save(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase = None ): '''simple docstring''' if resolved_file is None or commit_hash is not None: return commit_hash UpperCAmelCase__ : Any = str(Path(__UpperCamelCase ).as_posix() ) UpperCAmelCase__ : Union[str, Any] = re.search(r"""snapshots/([^/]+)/""" , __UpperCamelCase ) if search is None: return None UpperCAmelCase__ : Optional[Any] = search.groups()[0] return commit_hash if REGEX_COMMIT_HASH.match(__UpperCamelCase ) else None # Old default cache path, potentially to be migrated. # This logic was more or less taken from `transformers`, with the following differences: # - Diffusers doesn't use custom environment variables to specify the cache path. # - There is no need to migrate the cache format, just move the files to the new location. __UpperCAmelCase = os.path.expanduser( os.getenv('HF_HOME', os.path.join(os.getenv('XDG_CACHE_HOME', '~/.cache'), 'huggingface')) ) __UpperCAmelCase = os.path.join(hf_cache_home, 'diffusers') def lowerCAmelCase ( __UpperCamelCase = None , __UpperCamelCase = None ): '''simple docstring''' if new_cache_dir is None: UpperCAmelCase__ : Dict = DIFFUSERS_CACHE if old_cache_dir is None: UpperCAmelCase__ : Tuple = old_diffusers_cache UpperCAmelCase__ : List[Any] = Path(__UpperCamelCase ).expanduser() UpperCAmelCase__ : Optional[int] = Path(__UpperCamelCase ).expanduser() for old_blob_path in old_cache_dir.glob("""**/blobs/*""" ): if old_blob_path.is_file() and not old_blob_path.is_symlink(): UpperCAmelCase__ : Optional[Any] = new_cache_dir / old_blob_path.relative_to(__UpperCamelCase ) new_blob_path.parent.mkdir(parents=__UpperCamelCase , exist_ok=__UpperCamelCase ) os.replace(__UpperCamelCase , __UpperCamelCase ) try: os.symlink(__UpperCamelCase , __UpperCamelCase ) except OSError: logger.warning( """Could not create symlink between old cache and new cache. If you use an older version of diffusers again, files will be re-downloaded.""" ) # At this point, old_cache_dir contains symlinks to the new cache (it can still be used). __UpperCAmelCase = os.path.join(DIFFUSERS_CACHE, 'version_diffusers_cache.txt') if not os.path.isfile(cache_version_file): __UpperCAmelCase = 0 else: with open(cache_version_file) as f: try: __UpperCAmelCase = int(f.read()) except ValueError: __UpperCAmelCase = 0 if cache_version < 1: __UpperCAmelCase = os.path.isdir(old_diffusers_cache) and len(os.listdir(old_diffusers_cache)) > 0 if old_cache_is_not_empty: logger.warning( 'The cache for model files in Diffusers v0.14.0 has moved to a new location. Moving your ' 'existing cached models. This is a one-time operation, you can interrupt it or run it ' 'later by calling `diffusers.utils.hub_utils.move_cache()`.' ) try: move_cache() except Exception as e: __UpperCAmelCase = '\n'.join(traceback.format_tb(e.__traceback__)) logger.error( F"There was a problem when trying to move your cache:\n\n{trace}\n{e.__class__.__name__}: {e}\n\nPlease " 'file an issue at https://github.com/huggingface/diffusers/issues/new/choose, copy paste this whole ' 'message and we will do our best to help.' ) if cache_version < 1: try: os.makedirs(DIFFUSERS_CACHE, exist_ok=True) with open(cache_version_file, 'w') as f: f.write('1') except Exception: logger.warning( F"There was a problem when trying to write in your cache folder ({DIFFUSERS_CACHE}). Please, ensure " 'the directory exists and can be written to.' ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase = None ): '''simple docstring''' if variant is not None: UpperCAmelCase__ : int = weights_name.split(""".""" ) UpperCAmelCase__ : List[str] = splits[:-1] + [variant] + splits[-1:] UpperCAmelCase__ : List[Any] = """.""".join(__UpperCamelCase ) return weights_name def lowerCAmelCase ( __UpperCamelCase , *, __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , ): '''simple docstring''' UpperCAmelCase__ : Tuple = str(__UpperCamelCase ) if os.path.isfile(__UpperCamelCase ): return pretrained_model_name_or_path elif os.path.isdir(__UpperCamelCase ): if os.path.isfile(os.path.join(__UpperCamelCase , __UpperCamelCase ) ): # Load from a PyTorch checkpoint UpperCAmelCase__ : Tuple = os.path.join(__UpperCamelCase , __UpperCamelCase ) return model_file elif subfolder is not None and os.path.isfile( os.path.join(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) ): UpperCAmelCase__ : Optional[int] = os.path.join(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return model_file else: raise EnvironmentError( F"Error no file named {weights_name} found in directory {pretrained_model_name_or_path}." ) else: # 1. First check if deprecated way of loading from branches is used if ( revision in DEPRECATED_REVISION_ARGS and (weights_name == WEIGHTS_NAME or weights_name == SAFETENSORS_WEIGHTS_NAME) and version.parse(version.parse(__UpperCamelCase ).base_version ) >= version.parse("""0.20.0""" ) ): try: UpperCAmelCase__ : List[str] = hf_hub_download( __UpperCamelCase , filename=_add_variant(__UpperCamelCase , __UpperCamelCase ) , cache_dir=__UpperCamelCase , force_download=__UpperCamelCase , proxies=__UpperCamelCase , resume_download=__UpperCamelCase , local_files_only=__UpperCamelCase , use_auth_token=__UpperCamelCase , user_agent=__UpperCamelCase , subfolder=__UpperCamelCase , revision=revision or commit_hash , ) warnings.warn( F"Loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'` is deprecated. Loading instead from `revision='main'` with `variant={revision}`. Loading model variants via `revision='{revision}'` will be removed in diffusers v1. Please use `variant='{revision}'` instead." , __UpperCamelCase , ) return model_file except: # noqa: E722 warnings.warn( F"You are loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'`. This behavior is deprecated and will be removed in diffusers v1. One should use `variant='{revision}'` instead. However, it appears that {pretrained_model_name_or_path} currently does not have a {_add_variant(__UpperCamelCase , __UpperCamelCase )} file in the 'main' branch of {pretrained_model_name_or_path}. \n The Diffusers team and community would be very grateful if you could open an issue: https://github.com/huggingface/diffusers/issues/new with the title '{pretrained_model_name_or_path} is missing {_add_variant(__UpperCamelCase , __UpperCamelCase )}' so that the correct variant file can be added." , __UpperCamelCase , ) try: # 2. Load model file as usual UpperCAmelCase__ : List[Any] = hf_hub_download( __UpperCamelCase , filename=__UpperCamelCase , cache_dir=__UpperCamelCase , force_download=__UpperCamelCase , proxies=__UpperCamelCase , resume_download=__UpperCamelCase , local_files_only=__UpperCamelCase , use_auth_token=__UpperCamelCase , user_agent=__UpperCamelCase , subfolder=__UpperCamelCase , revision=revision or commit_hash , ) return model_file except RepositoryNotFoundError: raise EnvironmentError( F"{pretrained_model_name_or_path} is not a local folder and is not a valid model identifier " """listed on 'https://huggingface.co/models'\nIf this is a private repository, make sure to pass a """ """token having permission to this repo with `use_auth_token` or log in with `huggingface-cli """ """login`.""" ) except RevisionNotFoundError: raise EnvironmentError( F"{revision} is not a valid git identifier (branch name, tag name or commit id) that exists for " """this model name. Check the model page at """ F"'https://huggingface.co/{pretrained_model_name_or_path}' for available revisions." ) except EntryNotFoundError: raise EnvironmentError( F"{pretrained_model_name_or_path} does not appear to have a file named {weights_name}." ) except HTTPError as err: raise EnvironmentError( F"There was a specific connection error when trying to load {pretrained_model_name_or_path}:\n{err}" ) except ValueError: raise EnvironmentError( F"We couldn't connect to '{HUGGINGFACE_CO_RESOLVE_ENDPOINT}' to load this model, couldn't find it" F" in the cached files and it looks like {pretrained_model_name_or_path} is not the path to a" F" directory containing a file named {weights_name} or" """ \nCheckout your internet connection or see how to run the library in""" """ offline mode at 'https://huggingface.co/docs/diffusers/installation#offline-mode'.""" ) except EnvironmentError: raise EnvironmentError( F"Can't load the model for '{pretrained_model_name_or_path}'. If you were trying to load it from " """'https://huggingface.co/models', make sure you don't have a local directory with the same name. """ F"Otherwise, make sure '{pretrained_model_name_or_path}' is the correct path to a directory " F"containing a file named {weights_name}" )
65
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class __lowercase ( __lowerCamelCase ): snake_case_ = """open-llama""" def __init__( self : Dict ,A : str=100_000 ,A : str=4_096 ,A : Optional[Any]=11_008 ,A : Tuple=32 ,A : str=32 ,A : Optional[int]="silu" ,A : List[Any]=2_048 ,A : str=0.0_2 ,A : Optional[int]=1e-6 ,A : int=True ,A : Tuple=0 ,A : str=1 ,A : Any=2 ,A : Optional[Any]=False ,A : int=True ,A : Any=0.1 ,A : Optional[Any]=0.1 ,A : Optional[Any]=True ,A : Union[str, Any]=True ,A : Tuple=None ,**A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : str = vocab_size UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : Tuple = intermediate_size UpperCAmelCase__ : Optional[int] = num_hidden_layers UpperCAmelCase__ : Any = num_attention_heads UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = initializer_range UpperCAmelCase__ : Optional[int] = rms_norm_eps UpperCAmelCase__ : Any = use_cache UpperCAmelCase__ : Optional[Any] = kwargs.pop( """use_memorry_efficient_attention""" ,A ) UpperCAmelCase__ : Any = hidden_dropout_prob UpperCAmelCase__ : str = attention_dropout_prob UpperCAmelCase__ : Optional[int] = use_stable_embedding UpperCAmelCase__ : Tuple = shared_input_output_embedding UpperCAmelCase__ : Tuple = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,tie_word_embeddings=A ,**A ,) def __lowercase ( self : Optional[Any] ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ f"got {self.rope_scaling}" ) UpperCAmelCase__ : List[Any] = self.rope_scaling.get("""type""" ,A ) UpperCAmelCase__ : int = self.rope_scaling.get("""factor""" ,A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(A ,A ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
65
1
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Any = { """repo_name""": ["""test_repo1""", """test_repo2""", """test_repo3"""], """path""": ["""test_1.py""", """test_2.py""", """unit_test.py"""], """content""": ["""a """ * 20, """a """ * 30, """b """ * 7], } UpperCAmelCase__ : Union[str, Any] = Dataset.from_dict(__UpperCamelCase ) return dataset class __lowercase ( __lowerCamelCase ): def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Any = get_dataset() UpperCAmelCase__ : int = make_duplicate_clusters(A ,0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ) ,2 ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Tuple = get_dataset() UpperCAmelCase__ , UpperCAmelCase__ : Dict = deduplicate_dataset(A ) self.assertEqual(len(A ) ,2 ) print(A ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] ,2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] ,A )
65
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" from functools import reduce __UpperCAmelCase = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def lowerCAmelCase ( __UpperCamelCase = N ): '''simple docstring''' return max( # mypy cannot properly interpret reduce int(reduce(lambda __UpperCamelCase , __UpperCamelCase : str(int(__UpperCamelCase ) * int(__UpperCamelCase ) ) , n[i : i + 13] ) ) for i in range(len(__UpperCamelCase ) - 12 ) ) if __name__ == "__main__": print(F"{solution() = }")
65
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """attention_mask"""] def __init__( self : Any ,A : str=80 ,A : Optional[int]=16_000 ,A : int=0.0 ,A : str=10 ,A : Any=25 ,A : str="hamming_window" ,A : int=3_2_7_6_8.0 ,A : List[str]=0.9_7 ,A : Optional[int]=1.0 ,A : Optional[Any]=True ,A : Tuple=True ,A : Any=False ,**A : int ,): '''simple docstring''' super().__init__(feature_size=A ,sampling_rate=A ,padding_value=A ,**A ) UpperCAmelCase__ : str = feature_size UpperCAmelCase__ : int = sampling_rate UpperCAmelCase__ : int = padding_value UpperCAmelCase__ : Dict = hop_length UpperCAmelCase__ : int = win_length UpperCAmelCase__ : Dict = frame_signal_scale UpperCAmelCase__ : Dict = preemphasis_coeff UpperCAmelCase__ : str = mel_floor UpperCAmelCase__ : Any = normalize_means UpperCAmelCase__ : str = normalize_vars UpperCAmelCase__ : int = win_function UpperCAmelCase__ : List[Any] = return_attention_mask UpperCAmelCase__ : str = win_length * sampling_rate // 1_000 UpperCAmelCase__ : List[Any] = hop_length * sampling_rate // 1_000 UpperCAmelCase__ : int = optimal_fft_length(self.sample_size ) UpperCAmelCase__ : List[Any] = (self.n_fft // 2) + 1 def __lowercase ( self : Union[str, Any] ,A : np.array ): '''simple docstring''' if self.win_function == "hamming_window": UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=A ) else: UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ) UpperCAmelCase__ : Union[str, Any] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) UpperCAmelCase__ : Optional[Any] = spectrogram( one_waveform * self.frame_signal_scale ,window=A ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=A ,preemphasis=self.preemphasis_coeff ,mel_filters=A ,mel_floor=self.mel_floor ,log_mel="""log""" ,) return msfc_features.T def __lowercase ( self : str ,A : Any ,A : Optional[int] ,A : str ): '''simple docstring''' # make sure we normalize float32 arrays if self.normalize_means: UpperCAmelCase__ : Optional[Any] = x[:input_length].mean(axis=0 ) UpperCAmelCase__ : Any = np.subtract(A ,A ) if self.normalize_vars: UpperCAmelCase__ : str = x[:input_length].std(axis=0 ) UpperCAmelCase__ : Optional[int] = np.divide(A ,A ) if input_length < x.shape[0]: UpperCAmelCase__ : int = padding_value # make sure array is in float32 UpperCAmelCase__ : str = x.astype(np.floataa ) return x def __lowercase ( self : Union[str, Any] ,A : List[np.ndarray] ,A : Optional[np.ndarray] = None ): '''simple docstring''' UpperCAmelCase__ : Any = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(A ,A ,self.padding_value ) for x, n in zip(A ,A )] def __call__( self : Union[str, Any] ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : Union[bool, str, PaddingStrategy] = False ,A : Optional[int] = None ,A : bool = False ,A : Optional[int] = None ,A : Optional[bool] = None ,A : Optional[Union[str, TensorType]] = None ,A : Optional[int] = None ,**A : Tuple ,): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" f" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" f" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[Any] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : Any = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : List[str] = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Union[str, Any] = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [raw_speech] # extract fbank features UpperCAmelCase__ : Tuple = [self._extract_mfsc_features(A ) for one_waveform in raw_speech] # convert into correct format for padding UpperCAmelCase__ : str = BatchFeature({"""input_features""": features} ) UpperCAmelCase__ : Optional[Any] = self.pad( A ,padding=A ,max_length=A ,truncation=A ,pad_to_multiple_of=A ,return_attention_mask=A ,**A ,) # make sure list is in array format UpperCAmelCase__ : Tuple = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] ,A ): UpperCAmelCase__ : Union[str, Any] = [np.asarray(A ,dtype=np.floataa ) for feature in input_features] UpperCAmelCase__ : Dict = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: UpperCAmelCase__ : Union[str, Any] = ( np.array(A ,dtype=np.intaa ) if self._get_padding_strategies(A ,max_length=A ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) UpperCAmelCase__ : Any = self.normalize( padded_inputs["""input_features"""] ,attention_mask=A ) if return_tensors is not None: UpperCAmelCase__ : Union[str, Any] = padded_inputs.convert_to_tensors(A ) return padded_inputs
65
1
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = {'vocab_file': 'spiece.model'} __UpperCAmelCase = { 'vocab_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model', } } __UpperCAmelCase = { 'xlnet-base-cased': None, 'xlnet-large-cased': None, } # Segments (not really needed) __UpperCAmelCase = 0 __UpperCAmelCase = 1 __UpperCAmelCase = 2 __UpperCAmelCase = 3 __UpperCAmelCase = 4 class __lowercase ( __lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = """left""" def __init__( self : List[Any] ,A : Optional[int] ,A : List[Any]=False ,A : List[Any]=True ,A : Optional[int]=False ,A : Optional[int]="<s>" ,A : List[str]="</s>" ,A : List[str]="<unk>" ,A : Dict="<sep>" ,A : Dict="<pad>" ,A : Tuple="<cls>" ,A : str="<mask>" ,A : Any=["<eop>", "<eod>"] ,A : Optional[Dict[str, Any]] = None ,**A : Tuple ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ : Any = AddedToken(A ,lstrip=A ,rstrip=A ) if isinstance(A ,A ) else mask_token UpperCAmelCase__ : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=A ,remove_space=A ,keep_accents=A ,bos_token=A ,eos_token=A ,unk_token=A ,sep_token=A ,pad_token=A ,cls_token=A ,mask_token=A ,additional_special_tokens=A ,sp_model_kwargs=self.sp_model_kwargs ,**A ,) UpperCAmelCase__ : Dict = 3 UpperCAmelCase__ : List[str] = do_lower_case UpperCAmelCase__ : List[Any] = remove_space UpperCAmelCase__ : List[Any] = keep_accents UpperCAmelCase__ : Tuple = vocab_file UpperCAmelCase__ : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : int = {self.convert_ids_to_tokens(A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.__dict__.copy() UpperCAmelCase__ : Optional[Any] = None return state def __setstate__( self : Tuple ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Any = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): UpperCAmelCase__ : Dict = {} UpperCAmelCase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __lowercase ( self : int ,A : Dict ): '''simple docstring''' if self.remove_space: UpperCAmelCase__ : Union[str, Any] = """ """.join(inputs.strip().split() ) else: UpperCAmelCase__ : Tuple = inputs UpperCAmelCase__ : List[Any] = outputs.replace("""``""" ,"""\"""" ).replace("""''""" ,"""\"""" ) if not self.keep_accents: UpperCAmelCase__ : str = unicodedata.normalize("""NFKD""" ,A ) UpperCAmelCase__ : List[Any] = """""".join([c for c in outputs if not unicodedata.combining(A )] ) if self.do_lower_case: UpperCAmelCase__ : List[Any] = outputs.lower() return outputs def __lowercase ( self : Any ,A : str ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.preprocess_text(A ) UpperCAmelCase__ : Dict = self.sp_model.encode(A ,out_type=A ) UpperCAmelCase__ : Optional[Any] = [] for piece in pieces: if len(A ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): UpperCAmelCase__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(A ,"""""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: UpperCAmelCase__ : str = cur_pieces[1:] else: UpperCAmelCase__ : Union[str, Any] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(A ) else: new_pieces.append(A ) return new_pieces def __lowercase ( self : Any ,A : Optional[Any] ): '''simple docstring''' return self.sp_model.PieceToId(A ) def __lowercase ( self : Tuple ,A : Dict ): '''simple docstring''' return self.sp_model.IdToPiece(A ) def __lowercase ( self : Tuple ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """""".join(A ).replace(A ,""" """ ).strip() return out_string def __lowercase ( self : Tuple ,A : List[int] ,A : bool = False ,A : bool = None ,A : bool = True ,**A : List[str] ,): '''simple docstring''' UpperCAmelCase__ : str = kwargs.pop("""use_source_tokenizer""" ,A ) UpperCAmelCase__ : List[Any] = self.convert_ids_to_tokens(A ,skip_special_tokens=A ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 UpperCAmelCase__ : List[Any] = [] UpperCAmelCase__ : int = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(A ) ) UpperCAmelCase__ : Dict = [] sub_texts.append(A ) else: current_sub_text.append(A ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(A ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens UpperCAmelCase__ : str = """""".join(A ) UpperCAmelCase__ : Tuple = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: UpperCAmelCase__ : Any = self.clean_up_tokenization(A ) return clean_text else: return text def __lowercase ( self : Any ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = [self.sep_token_id] UpperCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def __lowercase ( self : str ,A : List[int] ,A : Optional[List[int]] = None ,A : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A ,token_ids_a=A ,already_has_special_tokens=A ) if token_ids_a is not None: return ([0] * len(A )) + [1] + ([0] * len(A )) + [1, 1] return ([0] * len(A )) + [1, 1] def __lowercase ( self : Optional[int] ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = [self.sep_token_id] UpperCAmelCase__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def __lowercase ( self : Union[str, Any] ,A : str ,A : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(A ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase__ : Any = os.path.join( A ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,A ) elif not os.path.isfile(self.vocab_file ): with open(A ,"""wb""" ) as fi: UpperCAmelCase__ : str = self.sp_model.serialized_model_proto() fi.write(A ) return (out_vocab_file,)
65
"""simple docstring""" from math import factorial def lowerCAmelCase ( __UpperCamelCase = 100 ): '''simple docstring''' return sum(int(__UpperCamelCase ) for x in str(factorial(__UpperCamelCase ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
65
1
"""simple docstring""" from ...processing_utils import ProcessorMixin class __lowercase ( __lowerCamelCase ): snake_case_ = """SpeechT5FeatureExtractor""" snake_case_ = """SpeechT5Tokenizer""" def __init__( self : Optional[int] ,A : str ,A : Any ): '''simple docstring''' super().__init__(A ,A ) def __call__( self : str ,*A : Tuple ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = kwargs.pop("""audio""" ,A ) UpperCAmelCase__ : Union[str, Any] = kwargs.pop("""text""" ,A ) UpperCAmelCase__ : Optional[int] = kwargs.pop("""text_target""" ,A ) UpperCAmelCase__ : Any = kwargs.pop("""audio_target""" ,A ) UpperCAmelCase__ : List[str] = kwargs.pop("""sampling_rate""" ,A ) if audio is not None and text is not None: raise ValueError( """Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?""" ) if audio_target is not None and text_target is not None: raise ValueError( """Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?""" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( """You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.""" ) if audio is not None: UpperCAmelCase__ : Union[str, Any] = self.feature_extractor(A ,*A ,sampling_rate=A ,**A ) elif text is not None: UpperCAmelCase__ : Optional[int] = self.tokenizer(A ,**A ) else: UpperCAmelCase__ : Optional[int] = None if audio_target is not None: UpperCAmelCase__ : List[Any] = self.feature_extractor(audio_target=A ,*A ,sampling_rate=A ,**A ) UpperCAmelCase__ : int = targets["""input_values"""] elif text_target is not None: UpperCAmelCase__ : List[str] = self.tokenizer(A ,**A ) UpperCAmelCase__ : Optional[Any] = targets["""input_ids"""] else: UpperCAmelCase__ : List[Any] = None if inputs is None: return targets if targets is not None: UpperCAmelCase__ : Optional[Any] = labels UpperCAmelCase__ : Any = targets.get("""attention_mask""" ) if decoder_attention_mask is not None: UpperCAmelCase__ : List[Any] = decoder_attention_mask return inputs def __lowercase ( self : Optional[int] ,*A : int ,**A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = kwargs.pop("""input_values""" ,A ) UpperCAmelCase__ : Tuple = kwargs.pop("""input_ids""" ,A ) UpperCAmelCase__ : Union[str, Any] = kwargs.pop("""labels""" ,A ) if input_values is not None and input_ids is not None: raise ValueError("""Cannot process both `input_values` and `input_ids` inputs.""" ) if input_values is None and input_ids is None and labels is None: raise ValueError( """You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.""" ) if input_values is not None: UpperCAmelCase__ : List[str] = self.feature_extractor.pad(A ,*A ,**A ) elif input_ids is not None: UpperCAmelCase__ : Dict = self.tokenizer.pad(A ,**A ) else: UpperCAmelCase__ : Dict = None if labels is not None: if "input_ids" in labels or (isinstance(A ,A ) and "input_ids" in labels[0]): UpperCAmelCase__ : int = self.tokenizer.pad(A ,**A ) UpperCAmelCase__ : Optional[Any] = targets["""input_ids"""] else: UpperCAmelCase__ : Union[str, Any] = self.feature_extractor.feature_size UpperCAmelCase__ : Optional[Any] = self.feature_extractor.num_mel_bins UpperCAmelCase__ : int = self.feature_extractor.pad(A ,*A ,**A ) UpperCAmelCase__ : List[str] = feature_size_hack UpperCAmelCase__ : Optional[int] = targets["""input_values"""] else: UpperCAmelCase__ : Optional[Any] = None if inputs is None: return targets if targets is not None: UpperCAmelCase__ : Any = labels UpperCAmelCase__ : Dict = targets.get("""attention_mask""" ) if decoder_attention_mask is not None: UpperCAmelCase__ : List[str] = decoder_attention_mask return inputs def __lowercase ( self : str ,*A : int ,**A : Any ): '''simple docstring''' return self.tokenizer.batch_decode(*A ,**A ) def __lowercase ( self : int ,*A : Any ,**A : Dict ): '''simple docstring''' return self.tokenizer.decode(*A ,**A )
65
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Optional[int] ,A : int=13 ,A : Tuple=7 ,A : Dict=True ,A : Optional[int]=True ,A : Tuple=True ,A : str=True ,A : Any=99 ,A : Tuple=32 ,A : Dict=5 ,A : Optional[int]=4 ,A : Dict=37 ,A : Any="gelu" ,A : Any=0.1 ,A : Optional[int]=0.1 ,A : Union[str, Any]=512 ,A : Any=16 ,A : List[str]=2 ,A : List[Any]=0.0_2 ,A : Optional[int]=4 ,): '''simple docstring''' UpperCAmelCase__ : Dict = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : Optional[int] = is_training UpperCAmelCase__ : Optional[Any] = use_attention_mask UpperCAmelCase__ : int = use_token_type_ids UpperCAmelCase__ : int = use_labels UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : List[Any] = type_vocab_size UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : List[Any] = num_choices def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : List[str] = None if self.use_attention_mask: UpperCAmelCase__ : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : int = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=A ,) return config, input_ids, attention_mask def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = config_and_inputs UpperCAmelCase__ : str = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = FlaxDistilBertModelTester(self ) @slow def __lowercase ( self : Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = model_class_name.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(A ) @require_flax class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) UpperCAmelCase__ : str = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ : Dict = model(A ,attention_mask=A )[0] UpperCAmelCase__ : List[Any] = (1, 11, 768) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Any = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,A ,atol=1e-4 ) )
65
1
"""simple docstring""" import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if gpta_config_file == "": UpperCAmelCase__ : Dict = GPTaConfig() else: UpperCAmelCase__ : Optional[int] = GPTaConfig.from_json_file(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = GPTaModel(__UpperCamelCase ) # Load weights from numpy load_tf_weights_in_gpta(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Save pytorch-model UpperCAmelCase__ : Dict = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME UpperCAmelCase__ : Any = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(F"Save PyTorch model to {pytorch_weights_dump_path}" ) torch.save(model.state_dict() , __UpperCamelCase ) print(F"Save configuration file to {pytorch_config_dump_path}" ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--gpt2_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--gpt2_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained OpenAI model. \n' 'This specifies the model architecture.' ), ) __UpperCAmelCase = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
65
"""simple docstring""" __UpperCAmelCase = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image', 'mask_image']) __UpperCAmelCase = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['example_image', 'image', 'mask_image']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset(['input_tokens']) __UpperCAmelCase = frozenset(['input_tokens'])
65
1
"""simple docstring""" import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = """linear""" snake_case_ = """cosine""" snake_case_ = """cosine_with_restarts""" snake_case_ = """polynomial""" snake_case_ = """constant""" snake_case_ = """constant_with_warmup""" snake_case_ = """piecewise_constant""" def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase = -1 ): '''simple docstring''' return LambdaLR(__UpperCamelCase , lambda __UpperCamelCase : 1 , last_epoch=__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = -1 ): '''simple docstring''' def lr_lambda(__UpperCamelCase ): if current_step < num_warmup_steps: return float(__UpperCamelCase ) / float(max(1.0 , __UpperCamelCase ) ) return 1.0 return LambdaLR(__UpperCamelCase , __UpperCamelCase , last_epoch=__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = -1 ): '''simple docstring''' UpperCAmelCase__ : Dict = {} UpperCAmelCase__ : Optional[int] = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: UpperCAmelCase__ , UpperCAmelCase__ : int = rule_str.split(""":""" ) UpperCAmelCase__ : Any = int(__UpperCamelCase ) UpperCAmelCase__ : Dict = float(__UpperCamelCase ) UpperCAmelCase__ : int = value UpperCAmelCase__ : Dict = float(rule_list[-1] ) def create_rules_function(__UpperCamelCase , __UpperCamelCase ): def rule_func(__UpperCamelCase ) -> float: UpperCAmelCase__ : Optional[Any] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__UpperCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func UpperCAmelCase__ : Dict = create_rules_function(__UpperCamelCase , __UpperCamelCase ) return LambdaLR(__UpperCamelCase , __UpperCamelCase , last_epoch=__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=-1 ): '''simple docstring''' def lr_lambda(__UpperCamelCase ): if current_step < num_warmup_steps: return float(__UpperCamelCase ) / float(max(1 , __UpperCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 0.5 , __UpperCamelCase = -1 ): '''simple docstring''' def lr_lambda(__UpperCamelCase ): if current_step < num_warmup_steps: return float(__UpperCamelCase ) / float(max(1 , __UpperCamelCase ) ) UpperCAmelCase__ : Optional[Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__UpperCamelCase ) * 2.0 * progress )) ) return LambdaLR(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 1 , __UpperCamelCase = -1 ): '''simple docstring''' def lr_lambda(__UpperCamelCase ): if current_step < num_warmup_steps: return float(__UpperCamelCase ) / float(max(1 , __UpperCamelCase ) ) UpperCAmelCase__ : Optional[Any] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__UpperCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=1e-7 , __UpperCamelCase=1.0 , __UpperCamelCase=-1 ): '''simple docstring''' UpperCAmelCase__ : List[Any] = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(F"lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})" ) def lr_lambda(__UpperCamelCase ): if current_step < num_warmup_steps: return float(__UpperCamelCase ) / float(max(1 , __UpperCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: UpperCAmelCase__ : List[str] = lr_init - lr_end UpperCAmelCase__ : List[Any] = num_training_steps - num_warmup_steps UpperCAmelCase__ : Union[str, Any] = 1 - (current_step - num_warmup_steps) / decay_steps UpperCAmelCase__ : Any = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) __UpperCAmelCase = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = 1 , __UpperCamelCase = 1.0 , __UpperCamelCase = -1 , ): '''simple docstring''' UpperCAmelCase__ : List[str] = SchedulerType(__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__UpperCamelCase , last_epoch=__UpperCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__UpperCamelCase , step_rules=__UpperCamelCase , last_epoch=__UpperCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"{name} requires `num_warmup_steps`, please provide that argument." ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__UpperCamelCase , num_warmup_steps=__UpperCamelCase , last_epoch=__UpperCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"{name} requires `num_training_steps`, please provide that argument." ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __UpperCamelCase , num_warmup_steps=__UpperCamelCase , num_training_steps=__UpperCamelCase , num_cycles=__UpperCamelCase , last_epoch=__UpperCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __UpperCamelCase , num_warmup_steps=__UpperCamelCase , num_training_steps=__UpperCamelCase , power=__UpperCamelCase , last_epoch=__UpperCamelCase , ) return schedule_func( __UpperCamelCase , num_warmup_steps=__UpperCamelCase , num_training_steps=__UpperCamelCase , last_epoch=__UpperCamelCase )
65
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase__ : Tuple = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[Any] = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase__ : int = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 16_000, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase__ : Optional[int] = tempfile.mkdtemp() UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,A ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.feature_extraction_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) # load decoder from hub UpperCAmelCase__ : int = """hf-internal-testing/ngram-beam-search-decoder""" def __lowercase ( self : str ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.add_kwargs_tokens_map.copy() kwargs.update(A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[str] ,**A : Dict ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Any ,**A : List[Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name ,**A ) def __lowercase ( self : Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_decoder() UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() ,feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor ,A ) # decoder self.assertEqual(processor.decoder._alphabet.labels ,decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set ,decoder.model_container[decoder._model_key]._unigram_set ,) self.assertIsInstance(processor.decoder ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname ,alpha=5.0 ,beta=3.0 ,score_boundary=-7.0 ,unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha ,5.0 ) self.assertEqual(processor.language_model.beta ,3.0 ) self.assertEqual(processor.language_model.score_boundary ,-7.0 ) self.assertEqual(processor.language_model.unk_score_offset ,3 ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(A ,"""include""" ): WavaVecaProcessorWithLM( tokenizer=A ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : Optional[Any] = feature_extractor(A ,return_tensors="""np""" ) UpperCAmelCase__ : List[Any] = processor(A ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : List[Any] = """This is a test string""" UpperCAmelCase__ : int = processor(text=A ) UpperCAmelCase__ : Dict = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __lowercase ( self : Tuple ,A : List[Any]=(2, 10, 16) ,A : Dict=77 ): '''simple docstring''' np.random.seed(A ) return np.random.rand(*A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits(shape=(10, 16) ,seed=13 ) UpperCAmelCase__ : Tuple = processor.decode(A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(A )[0] self.assertEqual(decoded_decoder[0] ,decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" ,decoded_processor.text ) self.assertEqual(decoded_decoder[-2] ,decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] ,decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def __lowercase ( self : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_feature_extractor() UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : List[str] = processor.batch_decode(A ) else: with get_context(A ).Pool() as pool: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(A ,A ) UpperCAmelCase__ : Optional[Any] = list(A ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A ,decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] ,decoded_processor.text ) self.assertListEqual(A ,decoded_processor.logit_score ) self.assertListEqual(A ,decoded_processor.lm_score ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits() UpperCAmelCase__ : Any = 15 UpperCAmelCase__ : Dict = -2_0.0 UpperCAmelCase__ : List[Any] = -4.0 UpperCAmelCase__ : Union[str, Any] = processor.batch_decode( A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[str] = decoded_processor_out.text UpperCAmelCase__ : List[str] = list(A ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( A ,A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] ,A ) self.assertTrue(np.array_equal(A ,decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] ,A ,atol=1e-3 ) ) self.assertTrue(np.array_equal(A ,decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] ,A ,atol=1e-3 ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits() UpperCAmelCase__ : Tuple = 2.0 UpperCAmelCase__ : str = 5.0 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = processor.batch_decode( A ,alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) UpperCAmelCase__ : Any = decoded_processor_out.text UpperCAmelCase__ : Union[str, Any] = list(A ) decoder.reset_params( alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch( A ,A ,) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] ,A ) UpperCAmelCase__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha ,2.0 ) self.assertEqual(lm_model.beta ,5.0 ) self.assertEqual(lm_model.unk_score_offset ,-2_0.0 ) self.assertEqual(lm_model.score_boundary ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : str = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Any = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[int] = os.listdir(A ) UpperCAmelCase__ : List[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(A ) UpperCAmelCase__ : Tuple = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[int] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Tuple = os.listdir(A ) UpperCAmelCase__ : Dict = os.listdir(A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = floats_list((3, 1_000) ) UpperCAmelCase__ : List[str] = processor_wavaveca(A ,return_tensors="""np""" ) UpperCAmelCase__ : Dict = processor_auto(A ,return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() ,input_auto[key].sum() ,delta=1e-2 ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Tuple = processor_wavaveca.batch_decode(A ) UpperCAmelCase__ : List[str] = processor_auto.batch_decode(A ) self.assertListEqual(decoded_wavaveca.text ,decoded_auto.text ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : int = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) self.assertListEqual( processor.model_input_names ,feature_extractor.model_input_names ,msg="""`processor` and `feature_extractor` model input names do not match""" ,) @staticmethod def __lowercase ( A : Optional[Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = self._get_dummy_logits()[0] UpperCAmelCase__ : List[str] = processor.decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ) ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""end_offset""" ) ,[1, 3, 5] ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : int = self._get_dummy_logits() UpperCAmelCase__ : Any = processor.batch_decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(A ,"""word""" ) ) for o in outputs["""word_offsets"""]] ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""end_offset""" ) ,[1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self : Tuple ): '''simple docstring''' import torch UpperCAmelCase__ : Any = load_dataset("""common_voice""" ,"""en""" ,split="""train""" ,streaming=A ) UpperCAmelCase__ : Tuple = ds.cast_column("""audio""" ,datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Tuple = iter(A ) UpperCAmelCase__ : Optional[int] = next(A ) UpperCAmelCase__ : List[Any] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase__ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Tuple = processor(sample["""audio"""]["""array"""] ,return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(A ).logits.cpu().numpy() UpperCAmelCase__ : Any = processor.decode(logits[0] ,output_word_offsets=A ) UpperCAmelCase__ : str = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase__ : Dict = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,A ) self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,output.text ) # output times UpperCAmelCase__ : str = torch.tensor(self.get_from_offsets(A ,"""start_time""" ) ) UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(A ,"""end_time""" ) ) # fmt: off UpperCAmelCase__ : Union[str, Any] = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : List[Any] = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) ) self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) )
65
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'microsoft/swin-tiny-patch4-window7-224': ( 'https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json' ), # See all Swin models at https://huggingface.co/models?filter=swin } class __lowercase ( __lowerCamelCase , __lowerCamelCase ): snake_case_ = """swin""" snake_case_ = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : List[str] ,A : List[str]=224 ,A : Tuple=4 ,A : str=3 ,A : Union[str, Any]=96 ,A : List[str]=[2, 2, 6, 2] ,A : List[Any]=[3, 6, 12, 24] ,A : List[Any]=7 ,A : List[str]=4.0 ,A : Optional[int]=True ,A : Optional[int]=0.0 ,A : List[str]=0.0 ,A : List[str]=0.1 ,A : Dict="gelu" ,A : Any=False ,A : List[str]=0.0_2 ,A : str=1e-5 ,A : Optional[Any]=32 ,A : List[str]=None ,A : Dict=None ,**A : int ,): '''simple docstring''' super().__init__(**A ) UpperCAmelCase__ : str = image_size UpperCAmelCase__ : str = patch_size UpperCAmelCase__ : Union[str, Any] = num_channels UpperCAmelCase__ : List[str] = embed_dim UpperCAmelCase__ : int = depths UpperCAmelCase__ : Dict = len(A ) UpperCAmelCase__ : Tuple = num_heads UpperCAmelCase__ : Any = window_size UpperCAmelCase__ : Any = mlp_ratio UpperCAmelCase__ : str = qkv_bias UpperCAmelCase__ : List[str] = hidden_dropout_prob UpperCAmelCase__ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase__ : int = drop_path_rate UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Any = use_absolute_embeddings UpperCAmelCase__ : Optional[Any] = layer_norm_eps UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : List[str] = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCAmelCase__ : Optional[Any] = int(embed_dim * 2 ** (len(A ) - 1) ) UpperCAmelCase__ : Optional[int] = ["""stem"""] + [f"stage{idx}" for idx in range(1 ,len(A ) + 1 )] UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = get_aligned_output_features_output_indices( out_features=A ,out_indices=A ,stage_names=self.stage_names ) class __lowercase ( __lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __lowercase ( self : Tuple ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' return 1e-4
65
"""simple docstring""" from sklearn.metrics import fa_score import datasets __UpperCAmelCase = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __UpperCAmelCase = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __UpperCAmelCase = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def __lowercase ( self : List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) ,reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"""] ,) def __lowercase ( self : Union[str, Any] ,A : List[str] ,A : List[Any] ,A : Optional[Any]=None ,A : List[str]=1 ,A : Optional[Any]="binary" ,A : Any=None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = fa_score( A ,A ,labels=A ,pos_label=A ,average=A ,sample_weight=A ) return {"f1": float(A ) if score.size == 1 else score}
65
1
"""simple docstring""" import os import sys __UpperCAmelCase = os.path.join(os.path.dirname(__file__), 'src') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) __UpperCAmelCase = [ 'torch', 'numpy', 'tokenizers', 'filelock', 'requests', 'tqdm', 'regex', 'sentencepiece', 'sacremoses', 'importlib_metadata', 'huggingface_hub', ] @add_start_docstrings(AutoConfig.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoConfig.from_pretrained(*__UpperCamelCase , **__UpperCamelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoTokenizer.from_pretrained(*__UpperCamelCase , **__UpperCamelCase ) @add_start_docstrings(AutoModel.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoModel.from_pretrained(*__UpperCamelCase , **__UpperCamelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoModelForCausalLM.from_pretrained(*__UpperCamelCase , **__UpperCamelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoModelForMaskedLM.from_pretrained(*__UpperCamelCase , **__UpperCamelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoModelForSequenceClassification.from_pretrained(*__UpperCamelCase , **__UpperCamelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def lowerCAmelCase ( *__UpperCamelCase , **__UpperCamelCase ): '''simple docstring''' return AutoModelForQuestionAnswering.from_pretrained(*__UpperCamelCase , **__UpperCamelCase )
65
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece.model') __UpperCAmelCase = {'target_lang': 'fi', 'source_lang': 'en'} __UpperCAmelCase = '>>zh<<' __UpperCAmelCase = 'Helsinki-NLP/' if is_torch_available(): __UpperCAmelCase = 'pt' elif is_tf_available(): __UpperCAmelCase = 'tf' else: __UpperCAmelCase = 'jax' @require_sentencepiece class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = MarianTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Optional[int] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Optional[Any] = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] UpperCAmelCase__ : int = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[int] = Path(self.tmpdirname ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(A ,save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(A ,save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) UpperCAmelCase__ : Dict = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self : List[Any] ,**A : List[Any] ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Union[str, Any] ,A : Tuple ): '''simple docstring''' return ( "This is a test", "This is a test", ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """</s>""" UpperCAmelCase__ : int = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""</s>""" ) self.assertEqual(vocab_keys[1] ,"""<unk>""" ) self.assertEqual(vocab_keys[-1] ,"""<pad>""" ) self.assertEqual(len(A ) ,9 ) def __lowercase ( self : Dict ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,9 ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MarianTokenizer.from_pretrained(f"{ORG_NAME}opus-mt-en-de" ) UpperCAmelCase__ : List[str] = en_de_tokenizer(["""I am a small frog"""] ,return_tensors=A ) self.assertIsInstance(A ,A ) UpperCAmelCase__ : str = [38, 121, 14, 697, 38_848, 0] self.assertListEqual(A ,batch.input_ids[0] ) UpperCAmelCase__ : Optional[Any] = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(A ) UpperCAmelCase__ : Tuple = [x.name for x in Path(A ).glob("""*""" )] self.assertIn("""source.spm""" ,A ) MarianTokenizer.from_pretrained(A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = tok( ["""I am a small frog""" * 1_000, """I am a small frog"""] ,padding=A ,truncation=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch.input_ids.shape ,(2, 512) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : Tuple = tok(["""I am a tiny frog""", """I am a small frog"""] ,padding=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch_smaller.input_ids.shape ,(2, 10) ) @slow def __lowercase ( self : Dict ): '''simple docstring''' # fmt: off UpperCAmelCase__ : Optional[int] = {"""input_ids""": [[43_495, 462, 20, 42_164, 1_369, 52, 464, 132, 1_703, 492, 13, 7_491, 38_999, 6, 8, 464, 132, 1_703, 492, 13, 4_669, 37_867, 13, 7_525, 27, 1_593, 988, 13, 33_972, 7_029, 6, 20, 8_251, 383, 2, 270, 5_866, 3_788, 2, 2_353, 8_251, 12_338, 2, 13_958, 387, 2, 3_629, 6_953, 188, 2_900, 2, 13_958, 8_011, 11_501, 23, 8_460, 4_073, 34_009, 20, 435, 11_439, 27, 8, 8_460, 4_073, 6_004, 20, 9_988, 375, 27, 33, 266, 1_945, 1_076, 1_350, 37_867, 3_288, 5, 577, 1_076, 4_374, 8, 5_082, 5, 26_453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10_767, 6, 316, 304, 4_239, 3, 0], [148, 15_722, 19, 1_839, 12, 1_350, 13, 22_327, 5_082, 5_418, 47_567, 35_938, 59, 318, 19_552, 108, 2_183, 54, 14_976, 4_835, 32, 547, 1_114, 8, 315, 2_417, 5, 92, 19_088, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100], [36, 6_395, 12_570, 39_147, 11_597, 6, 266, 4, 45_405, 7_296, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""Helsinki-NLP/opus-mt-en-de""" ,revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" ,decode_kwargs={"""use_source_tokenizer""": True} ,) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) UpperCAmelCase__ : Any = """Tämä on testi""" UpperCAmelCase__ : int = """This is a test""" UpperCAmelCase__ : List[str] = [76, 7, 2_047, 2] UpperCAmelCase__ : Optional[Any] = [69, 12, 11, 940, 2] UpperCAmelCase__ : List[str] = tokenizer(A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : Optional[int] = tokenizer(text_target=A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : int = tokenizer.decode(A ,skip_special_tokens=A ) self.assertEqual(A ,A )
65
1
"""simple docstring""" import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = BertJapaneseTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Any ): '''simple docstring''' super().setUp() UpperCAmelCase__ : int = [ """[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは""", """世界""", """##世界""", """、""", """##、""", """。""", """##。""", ] UpperCAmelCase__ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def __lowercase ( self : Dict ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """こんにちは、世界。 \nこんばんは、世界。""" UpperCAmelCase__ : List[str] = """こんにちは 、 世界 。 こんばんは 、 世界 。""" return input_text, output_text def __lowercase ( self : Tuple ,A : Tuple ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.get_input_output_texts(A ) UpperCAmelCase__ : List[str] = tokenizer.encode(A ,add_special_tokens=A ) UpperCAmelCase__ : Optional[Any] = tokenizer.decode(A ,clean_up_tokenization_spaces=A ) return text, ids def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass # TODO add if relevant def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass # TODO add if relevant def __lowercase ( self : str ): '''simple docstring''' pass # TODO add if relevant def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.tokenizer_class(self.vocab_file ) UpperCAmelCase__ : Any = tokenizer.tokenize("""こんにちは、世界。\nこんばんは、世界。""" ) self.assertListEqual(A ,["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class(self.vocab_file ,word_tokenizer_type="""mecab""" ) self.assertIsNotNone(A ) UpperCAmelCase__ : int = """こんにちは、世界。\nこんばんは、世界。""" UpperCAmelCase__ : Tuple = tokenizer.tokenize(A ) self.assertListEqual(A ,["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) UpperCAmelCase__ : List[Any] = os.path.join(self.tmpdirname ,"""tokenizer.bin""" ) with open(A ,"""wb""" ) as handle: pickle.dump(A ,A ) with open(A ,"""rb""" ) as handle: UpperCAmelCase__ : Tuple = pickle.load(A ) UpperCAmelCase__ : Optional[Any] = tokenizer_new.tokenize(A ) self.assertListEqual(A ,A ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = MecabTokenizer(mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] ,) def __lowercase ( self : Optional[Any] ): '''simple docstring''' try: UpperCAmelCase__ : Any = MecabTokenizer(mecab_dic="""unidic_lite""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] ,) def __lowercase ( self : str ): '''simple docstring''' try: UpperCAmelCase__ : Optional[Any] = MecabTokenizer(mecab_dic="""unidic""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] ,) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MecabTokenizer(do_lower_case=A ,mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップルストア""", """で""", """iphone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' try: UpperCAmelCase__ : Tuple = MecabTokenizer( do_lower_case=A ,normalize_text=A ,mecab_option="""-d /usr/local/lib/mecab/dic/jumandic""" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """\u3000""", """。"""] ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MecabTokenizer(normalize_text=A ,mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """ """, """。"""] ,) @require_sudachi def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.tokenizer_class(self.vocab_file ,word_tokenizer_type="""sudachi""" ) self.assertIsNotNone(A ) UpperCAmelCase__ : Optional[Any] = """こんにちは、世界。\nこんばんは、世界。""" UpperCAmelCase__ : Any = tokenizer.tokenize(A ) self.assertListEqual(A ,["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) UpperCAmelCase__ : List[Any] = os.path.join(self.tmpdirname ,"""tokenizer.bin""" ) with open(A ,"""wb""" ) as handle: pickle.dump(A ,A ) with open(A ,"""rb""" ) as handle: UpperCAmelCase__ : int = pickle.load(A ) UpperCAmelCase__ : Any = tokenizer_new.tokenize(A ) self.assertListEqual(A ,A ) @require_sudachi def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : str = SudachiTokenizer(sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,[""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] ,) @require_sudachi def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : str = SudachiTokenizer(sudachi_dict_type="""core""" ,sudachi_split_mode="""A""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) ,["""外国""", """人""", """参政""", """権"""] ) @require_sudachi def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Dict = SudachiTokenizer(sudachi_dict_type="""core""" ,sudachi_split_mode="""B""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) ,["""外国人""", """参政権"""] ) @require_sudachi def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = SudachiTokenizer(sudachi_dict_type="""core""" ,sudachi_split_mode="""C""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) ,["""外国人参政権"""] ) @require_sudachi def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : int = SudachiTokenizer(do_lower_case=A ,sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,[""" """, """\t""", """アップル""", """ストア""", """で""", """iphone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] ,) @require_sudachi def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = SudachiTokenizer(normalize_text=A ,sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,[""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """\u3000""", """。""", """ """, """ """] ,) @require_sudachi def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = SudachiTokenizer(trim_whitespace=A ,sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] ,) @require_jumanpp def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Dict = self.tokenizer_class(self.vocab_file ,word_tokenizer_type="""jumanpp""" ) self.assertIsNotNone(A ) UpperCAmelCase__ : Tuple = """こんにちは、世界。\nこんばんは、世界。""" UpperCAmelCase__ : Union[str, Any] = tokenizer.tokenize(A ) self.assertListEqual(A ,["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[3, 12, 10, 14, 4, 9, 12, 10, 14] ) UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,"""tokenizer.bin""" ) with open(A ,"""wb""" ) as handle: pickle.dump(A ,A ) with open(A ,"""rb""" ) as handle: UpperCAmelCase__ : Optional[int] = pickle.load(A ) UpperCAmelCase__ : Dict = tokenizer_new.tokenize(A ) self.assertListEqual(A ,A ) @require_jumanpp def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Tuple = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] ,) @require_jumanpp def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = JumanppTokenizer(do_lower_case=A ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップル""", """ストア""", """で""", """iphone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] ,) @require_jumanpp def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = JumanppTokenizer(normalize_text=A ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""ア""", """ッ""", """フ""", """゚""", """ル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] ,) @require_jumanpp def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : int = JumanppTokenizer(trim_whitespace=A ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) ,["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """。"""] ,) @require_jumanpp def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : List[Any] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("""ありがとうございますm(_ _)m見つけるのが大変です。""" ) ,["""ありがとう""", """ございます""", """m(_ _)m""", """見つける""", """の""", """が""", """大変です""", """。"""] ,) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは"""] UpperCAmelCase__ : int = {} for i, token in enumerate(A ): UpperCAmelCase__ : List[str] = i UpperCAmelCase__ : List[Any] = WordpieceTokenizer(vocab=A ,unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) ,[] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) ,["""こんにちは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは""" ) ,["""こん""", """##ばんは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは こんばんにちは こんにちは""" ) ,["""こん""", """##ばんは""", """[UNK]""", """こんにちは"""] ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : str = BertJapaneseTokenizer.from_pretrained("""nlp-waseda/roberta-base-japanese-with-auto-jumanpp""" ) UpperCAmelCase__ : Any = tokenizer.subword_tokenizer UpperCAmelCase__ : List[str] = subword_tokenizer.tokenize("""国境 の 長い トンネル を 抜ける と 雪国 であった 。""" ) self.assertListEqual(A ,["""▁国境""", """▁の""", """▁長い""", """▁トンネル""", """▁を""", """▁抜ける""", """▁と""", """▁雪""", """国""", """▁であった""", """▁。"""] ) UpperCAmelCase__ : str = subword_tokenizer.tokenize("""こんばんは こんばん にち は こんにちは""" ) self.assertListEqual(A ,["""▁こん""", """ばん""", """は""", """▁こん""", """ばん""", """▁に""", """ち""", """▁は""", """▁こんにちは"""] ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese""" ) UpperCAmelCase__ : Tuple = tokenizer.encode("""ありがとう。""" ,add_special_tokens=A ) UpperCAmelCase__ : int = tokenizer.encode("""どういたしまして。""" ,add_special_tokens=A ) UpperCAmelCase__ : Any = tokenizer.build_inputs_with_special_tokens(A ) UpperCAmelCase__ : str = tokenizer.build_inputs_with_special_tokens(A ,A ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = BertJapaneseTokenizer snake_case_ = False def __lowercase ( self : str ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Optional[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] UpperCAmelCase__ : int = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def __lowercase ( self : List[Any] ,**A : Optional[Any] ): '''simple docstring''' return BertJapaneseTokenizer.from_pretrained(self.tmpdirname ,subword_tokenizer_type="""character""" ,**A ) def __lowercase ( self : Any ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Any = """こんにちは、世界。 \nこんばんは、世界。""" UpperCAmelCase__ : Optional[int] = """こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。""" return input_text, output_text def __lowercase ( self : Any ): '''simple docstring''' pass # TODO add if relevant def __lowercase ( self : str ): '''simple docstring''' pass # TODO add if relevant def __lowercase ( self : List[Any] ): '''simple docstring''' pass # TODO add if relevant def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class(self.vocab_file ,subword_tokenizer_type="""character""" ) UpperCAmelCase__ : str = tokenizer.tokenize("""こんにちは、世界。 \nこんばんは、世界。""" ) self.assertListEqual( A ,["""こ""", """ん""", """に""", """ち""", """は""", """、""", """世""", """界""", """。""", """こ""", """ん""", """ば""", """ん""", """は""", """、""", """世""", """界""", """。"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A ) ,[3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : List[str] = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] UpperCAmelCase__ : Optional[int] = {} for i, token in enumerate(A ): UpperCAmelCase__ : List[Any] = i UpperCAmelCase__ : List[Any] = CharacterTokenizer(vocab=A ,unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) ,[] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) ,["""こ""", """ん""", """に""", """ち""", """は"""] ) self.assertListEqual(tokenizer.tokenize("""こんにちほ""" ) ,["""こ""", """ん""", """に""", """ち""", """[UNK]"""] ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese-char""" ) UpperCAmelCase__ : int = tokenizer.encode("""ありがとう。""" ,add_special_tokens=A ) UpperCAmelCase__ : Dict = tokenizer.encode("""どういたしまして。""" ,add_special_tokens=A ) UpperCAmelCase__ : Any = tokenizer.build_inputs_with_special_tokens(A ) UpperCAmelCase__ : Dict = tokenizer.build_inputs_with_special_tokens(A ,A ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowercase ( unittest.TestCase ): def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Dict = """cl-tohoku/bert-base-japanese""" UpperCAmelCase__ : str = AutoTokenizer.from_pretrained(A ) self.assertIsInstance(A ,A ) class __lowercase ( unittest.TestCase ): def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = """cl-tohoku/bert-base-japanese""" with self.assertLogs("""transformers""" ,level="""WARNING""" ) as cm: BertTokenizer.from_pretrained(A ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) ) UpperCAmelCase__ : Dict = """bert-base-cased""" with self.assertLogs("""transformers""" ,level="""WARNING""" ) as cm: BertJapaneseTokenizer.from_pretrained(A ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) )
65
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=__lowerCamelCase ): snake_case_ = ["""onnx"""] def __init__( self : int ,*A : List[str] ,**A : int ): '''simple docstring''' requires_backends(self ,["""onnx"""] ) @classmethod def __lowercase ( cls : Optional[Any] ,*A : List[str] ,**A : Dict ): '''simple docstring''' requires_backends(cls ,["""onnx"""] ) @classmethod def __lowercase ( cls : List[Any] ,*A : Optional[int] ,**A : int ): '''simple docstring''' requires_backends(cls ,["""onnx"""] )
65
1
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = {'configuration_focalnet': ['FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FocalNetConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'FocalNetForImageClassification', 'FocalNetForMaskedImageModeling', 'FocalNetBackbone', 'FocalNetModel', 'FocalNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) UpperCAmelCase__ : List[Any] = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(__UpperCamelCase ) # Let's go UpperCAmelCase__ : int = parser.parse_args() if not hasattr(__UpperCamelCase , """func""" ): parser.print_help() exit(1 ) # Run UpperCAmelCase__ : Union[str, Any] = args.func(__UpperCamelCase ) service.run() if __name__ == "__main__": main()
65
1
"""simple docstring""" import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = LxmertTokenizer snake_case_ = LxmertTokenizerFast snake_case_ = True snake_case_ = True def __lowercase ( self : Optional[Any] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : int = [ """[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] UpperCAmelCase__ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def __lowercase ( self : Union[str, Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Tuple = """UNwant\u00E9d,running""" UpperCAmelCase__ : int = """unwanted, running""" return input_text, output_text def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = self.tokenizer_class(self.vocab_file ) UpperCAmelCase__ : List[Any] = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(A ,["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,[7, 4, 5, 10, 8, 9] ) def __lowercase ( self : Any ): '''simple docstring''' if not self.test_rust_tokenizer: return UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_rust_tokenizer() UpperCAmelCase__ : Union[str, Any] = """I was born in 92000, and this is falsé.""" UpperCAmelCase__ : List[Any] = tokenizer.tokenize(A ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.tokenize(A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Dict = tokenizer.encode(A ,add_special_tokens=A ) UpperCAmelCase__ : int = rust_tokenizer.encode(A ,add_special_tokens=A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Any = self.get_rust_tokenizer() UpperCAmelCase__ : int = tokenizer.encode(A ) UpperCAmelCase__ : Any = rust_tokenizer.encode(A ) self.assertListEqual(A ,A )
65
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __UpperCAmelCase = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __lowercase : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : List[Any] ,A : int ,A : Optional[Any]=13 ,A : Dict=7 ,A : Dict=True ,A : Any=False ,A : Dict=99 ,A : int=32 ,A : Optional[int]=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : str=0.1 ,A : int=0.1 ,A : Optional[int]=20 ,A : Tuple=2 ,A : str=1 ,A : Optional[Any]=0 ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : int = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Union[str, Any] = pad_token_id UpperCAmelCase__ : List[str] = bos_token_id def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] ,self.vocab_size ).clip(3 ,self.vocab_size ) UpperCAmelCase__ : List[str] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) ,1 ) UpperCAmelCase__ : Any = np.concatenate([input_ids, eos_tensor] ,axis=1 ) UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : str = self.config_cls( vocab_size=self.vocab_size ,d_model=self.hidden_size ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,eos_token_ids=[2] ,bos_token_id=self.bos_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.pad_token_id ,**self.config_updates ,) UpperCAmelCase__ : Optional[Any] = prepare_pegasus_inputs_dict(A ,A ,A ) return config, inputs_dict def __lowercase ( self : Any ,A : Optional[int] ,A : str ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : Dict = model_class_name(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) ,dtype="""i4""" ) UpperCAmelCase__ : str = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Optional[int] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : int = model.decode( decoder_input_ids[:, -1:] ,A ,decoder_attention_mask=A ,past_key_values=outputs_cache.past_key_values ,decoder_position_ids=A ,) UpperCAmelCase__ : Dict = model.decode(A ,A ) UpperCAmelCase__ : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def __lowercase ( self : Optional[int] ,A : str ,A : Dict ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : str = model_class_name(A ) UpperCAmelCase__ : Any = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] ,axis=-1 ,) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : Dict = model.decode( decoder_input_ids[:, -1:] ,A ,past_key_values=outputs_cache.past_key_values ,decoder_attention_mask=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Union[str, Any] = model.decode(A ,A ,decoder_attention_mask=A ) UpperCAmelCase__ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , ): '''simple docstring''' if attention_mask is None: UpperCAmelCase__ : Union[str, Any] = np.not_equal(__UpperCamelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Tuple = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) snake_case_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = FlaxPegasusModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ) def __lowercase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(A ,A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(A ,A ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : List[Any] = self._prepare_for_class(A ,A ) UpperCAmelCase__ : int = model_class(A ) @jax.jit def encode_jitted(A : Optional[int] ,A : Union[str, Any]=None ,**A : Optional[Any] ): return model.encode(input_ids=A ,attention_mask=A ) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : int = encode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : Dict = encode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : Dict = model_class(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ,inputs_dict["""attention_mask"""] ) UpperCAmelCase__ : Dict = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(A : List[Any] ,A : Any ,A : List[Any] ): return model.decode( decoder_input_ids=A ,decoder_attention_mask=A ,encoder_outputs=A ,) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : Tuple = decode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : str = decode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""google/pegasus-large""" ,from_pt=A ) UpperCAmelCase__ : Any = np.ones((1, 1) ) UpperCAmelCase__ : Optional[Any] = model(A ) self.assertIsNotNone(A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Optional[Any] = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Union[str, Any] = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] UpperCAmelCase__ : str = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] UpperCAmelCase__ : str = tokenizer(A ,return_tensors="""np""" ,truncation=A ,max_length=512 ,padding=A ) UpperCAmelCase__ : Union[str, Any] = model.generate(**A ,num_beams=2 ).sequences UpperCAmelCase__ : int = tokenizer.batch_decode(A ,skip_special_tokens=A ) assert tgt_text == decoded
65
1
"""simple docstring""" from math import pi, sqrt, tan def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if side_length < 0: raise ValueError("""surface_area_cube() only accepts non-negative values""" ) return 6 * side_length**2 def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError("""surface_area_cuboid() only accepts non-negative values""" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if radius < 0: raise ValueError("""surface_area_sphere() only accepts non-negative values""" ) return 4 * pi * radius**2 def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if radius < 0: raise ValueError("""surface_area_hemisphere() only accepts non-negative values""" ) return 3 * pi * radius**2 def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if radius < 0 or height < 0: raise ValueError("""surface_area_cone() only accepts non-negative values""" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( """surface_area_conical_frustum() only accepts non-negative values""" ) UpperCAmelCase__ : Any = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if radius < 0 or height < 0: raise ValueError("""surface_area_cylinder() only accepts non-negative values""" ) return 2 * pi * radius * (height + radius) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError("""surface_area_torus() only accepts non-negative values""" ) if torus_radius < tube_radius: raise ValueError( """surface_area_torus() does not support spindle or self intersecting tori""" ) return 4 * pow(__UpperCamelCase , 2 ) * torus_radius * tube_radius def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if length < 0 or width < 0: raise ValueError("""area_rectangle() only accepts non-negative values""" ) return length * width def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if side_length < 0: raise ValueError("""area_square() only accepts non-negative values""" ) return side_length**2 def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if base < 0 or height < 0: raise ValueError("""area_triangle() only accepts non-negative values""" ) return (base * height) / 2 def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("""area_triangle_three_sides() only accepts non-negative values""" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("""Given three sides do not form a triangle""" ) UpperCAmelCase__ : List[str] = (sidea + sidea + sidea) / 2 UpperCAmelCase__ : Tuple = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if base < 0 or height < 0: raise ValueError("""area_parallelogram() only accepts non-negative values""" ) return base * height def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError("""area_trapezium() only accepts non-negative values""" ) return 1 / 2 * (basea + basea) * height def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if radius < 0: raise ValueError("""area_circle() only accepts non-negative values""" ) return pi * radius**2 def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError("""area_ellipse() only accepts non-negative values""" ) return pi * radius_x * radius_y def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError("""area_rhombus() only accepts non-negative values""" ) return 1 / 2 * diagonal_a * diagonal_a def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or sides < 3: raise ValueError( """area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides""" ) elif length < 0: raise ValueError( """area_reg_polygon() only accepts non-negative values as \ length of a side""" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(F"Rectangle: {area_rectangle(10, 20) = }") print(F"Square: {area_square(10) = }") print(F"Triangle: {area_triangle(10, 10) = }") print(F"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(F"Parallelogram: {area_parallelogram(10, 20) = }") print(F"Rhombus: {area_rhombus(10, 20) = }") print(F"Trapezium: {area_trapezium(10, 20, 30) = }") print(F"Circle: {area_circle(20) = }") print(F"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(F"Cube: {surface_area_cube(20) = }") print(F"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(F"Sphere: {surface_area_sphere(20) = }") print(F"Hemisphere: {surface_area_hemisphere(20) = }") print(F"Cone: {surface_area_cone(10, 20) = }") print(F"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(F"Cylinder: {surface_area_cylinder(10, 20) = }") print(F"Torus: {surface_area_torus(20, 10) = }") print(F"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(F"Square: {area_reg_polygon(4, 10) = }") print(F"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
65
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) UpperCAmelCase__ : Union[str, Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(__UpperCamelCase , n - 1 , __UpperCamelCase ) * a) % mod else: UpperCAmelCase__ : Union[str, Any] = binary_exponentiation(__UpperCamelCase , n / 2 , __UpperCamelCase ) return (b * b) % mod # a prime number __UpperCAmelCase = 701 __UpperCAmelCase = 10_0000_0000 __UpperCAmelCase = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
65
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase ( __UpperCamelCase = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCAmelCase__ : Dict = F"{olid} is not a valid Open Library olid" raise ValueError(__UpperCamelCase ) return requests.get(F"https://openlibrary.org/{new_olid}.json" ).json() def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } UpperCAmelCase__ : Dict = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase__ : str = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] UpperCAmelCase__ : Dict = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Dict = """, """.join(__UpperCamelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __UpperCAmelCase = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: __UpperCAmelCase = summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('\n'.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
65
1
"""simple docstring""" import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class __lowercase : @staticmethod def __lowercase ( *A : Any ,**A : Union[str, Any] ): '''simple docstring''' pass def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = np.array(__UpperCamelCase ) UpperCAmelCase__ : List[Any] = npimg.shape return {"hash": hashimage(__UpperCamelCase ), "shape": shape} @is_pipeline_test @require_vision @require_torch class __lowercase ( unittest.TestCase ): snake_case_ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) snake_case_ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def __lowercase ( self : Optional[Any] ,A : int ,A : Union[str, Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = MaskGenerationPipeline(model=A ,image_processor=A ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def __lowercase ( self : Optional[int] ,A : List[str] ,A : Dict ): '''simple docstring''' pass @require_tf @unittest.skip("""Image segmentation not implemented in TF""" ) def __lowercase ( self : List[str] ): '''simple docstring''' pass @slow @require_torch def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = pipeline("""mask-generation""" ,model="""facebook/sam-vit-huge""" ) UpperCAmelCase__ : Optional[Any] = image_segmenter("""http://images.cocodataset.org/val2017/000000039769.jpg""" ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ : Optional[int] = [] for i, o in enumerate(outputs["""masks"""] ): new_outupt += [{"mask": mask_to_test_readable(A ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(A ,decimals=4 ) ,[ {"""mask""": {"""hash""": """115ad19f5f""", """shape""": (480, 640)}, """scores""": 1.0_4_4_4}, {"""mask""": {"""hash""": """6affa964c6""", """shape""": (480, 640)}, """scores""": 1.0_2_1}, {"""mask""": {"""hash""": """dfe28a0388""", """shape""": (480, 640)}, """scores""": 1.0_1_6_7}, {"""mask""": {"""hash""": """c0a5f4a318""", """shape""": (480, 640)}, """scores""": 1.0_1_3_2}, {"""mask""": {"""hash""": """fe8065c197""", """shape""": (480, 640)}, """scores""": 1.0_0_5_3}, {"""mask""": {"""hash""": """e2d0b7a0b7""", """shape""": (480, 640)}, """scores""": 0.9_9_6_7}, {"""mask""": {"""hash""": """453c7844bd""", """shape""": (480, 640)}, """scores""": 0.9_9_3}, {"""mask""": {"""hash""": """3d44f2926d""", """shape""": (480, 640)}, """scores""": 0.9_9_0_9}, {"""mask""": {"""hash""": """64033ddc3f""", """shape""": (480, 640)}, """scores""": 0.9_8_7_9}, {"""mask""": {"""hash""": """801064ff79""", """shape""": (480, 640)}, """scores""": 0.9_8_3_4}, {"""mask""": {"""hash""": """6172f276ef""", """shape""": (480, 640)}, """scores""": 0.9_7_1_6}, {"""mask""": {"""hash""": """b49e60e084""", """shape""": (480, 640)}, """scores""": 0.9_6_1_2}, {"""mask""": {"""hash""": """a811e775fd""", """shape""": (480, 640)}, """scores""": 0.9_5_9_9}, {"""mask""": {"""hash""": """a6a8ebcf4b""", """shape""": (480, 640)}, """scores""": 0.9_5_5_2}, {"""mask""": {"""hash""": """9d8257e080""", """shape""": (480, 640)}, """scores""": 0.9_5_3_2}, {"""mask""": {"""hash""": """32de6454a8""", """shape""": (480, 640)}, """scores""": 0.9_5_1_6}, {"""mask""": {"""hash""": """af3d4af2c8""", """shape""": (480, 640)}, """scores""": 0.9_4_9_9}, {"""mask""": {"""hash""": """3c6db475fb""", """shape""": (480, 640)}, """scores""": 0.9_4_8_3}, {"""mask""": {"""hash""": """c290813fb9""", """shape""": (480, 640)}, """scores""": 0.9_4_6_4}, {"""mask""": {"""hash""": """b6f0b8f606""", """shape""": (480, 640)}, """scores""": 0.9_4_3}, {"""mask""": {"""hash""": """92ce16bfdf""", """shape""": (480, 640)}, """scores""": 0.9_4_3}, {"""mask""": {"""hash""": """c749b25868""", """shape""": (480, 640)}, """scores""": 0.9_4_0_8}, {"""mask""": {"""hash""": """efb6cab859""", """shape""": (480, 640)}, """scores""": 0.9_3_3_5}, {"""mask""": {"""hash""": """1ff2eafb30""", """shape""": (480, 640)}, """scores""": 0.9_3_2_6}, {"""mask""": {"""hash""": """788b798e24""", """shape""": (480, 640)}, """scores""": 0.9_2_6_2}, {"""mask""": {"""hash""": """abea804f0e""", """shape""": (480, 640)}, """scores""": 0.8_9_9_9}, {"""mask""": {"""hash""": """7b9e8ddb73""", """shape""": (480, 640)}, """scores""": 0.8_9_8_6}, {"""mask""": {"""hash""": """cd24047c8a""", """shape""": (480, 640)}, """scores""": 0.8_9_8_4}, {"""mask""": {"""hash""": """6943e6bcbd""", """shape""": (480, 640)}, """scores""": 0.8_8_7_3}, {"""mask""": {"""hash""": """b5f47c9191""", """shape""": (480, 640)}, """scores""": 0.8_8_7_1} ] ,) # fmt: on @require_torch @slow def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = """facebook/sam-vit-huge""" UpperCAmelCase__ : Dict = pipeline("""mask-generation""" ,model=A ) UpperCAmelCase__ : int = image_segmenter( """http://images.cocodataset.org/val2017/000000039769.jpg""" ,pred_iou_thresh=1 ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ : int = [] for i, o in enumerate(outputs["""masks"""] ): new_outupt += [{"mask": mask_to_test_readable(A ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(A ,decimals=4 ) ,[ {"""mask""": {"""hash""": """115ad19f5f""", """shape""": (480, 640)}, """scores""": 1.0_4_4_4}, {"""mask""": {"""hash""": """6affa964c6""", """shape""": (480, 640)}, """scores""": 1.0_2_1_0}, {"""mask""": {"""hash""": """dfe28a0388""", """shape""": (480, 640)}, """scores""": 1.0_1_6_7}, {"""mask""": {"""hash""": """c0a5f4a318""", """shape""": (480, 640)}, """scores""": 1.0_1_3_2}, {"""mask""": {"""hash""": """fe8065c197""", """shape""": (480, 640)}, """scores""": 1.0_0_5_3}, ] ,)
65
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=True , __UpperCamelCase="pt" ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {"""add_prefix_space""": True} if isinstance(__UpperCamelCase , __UpperCamelCase ) and not line.startswith(""" """ ) else {} UpperCAmelCase__ : List[str] = padding_side return tokenizer( [line] , max_length=__UpperCamelCase , padding="""max_length""" if pad_to_max_length else None , truncation=__UpperCamelCase , return_tensors=__UpperCamelCase , add_special_tokens=__UpperCamelCase , **__UpperCamelCase , ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , ): '''simple docstring''' UpperCAmelCase__ : str = input_ids.ne(__UpperCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ,A : List[Any] ,A : Union[str, Any] ,A : Any ,A : Optional[int] ,A : Union[str, Any]="train" ,A : Tuple=None ,A : Union[str, Any]=None ,A : Tuple=None ,A : int="" ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : Optional[Any] = Path(A ).joinpath(type_path + """.source""" ) UpperCAmelCase__ : List[str] = Path(A ).joinpath(type_path + """.target""" ) UpperCAmelCase__ : Dict = self.get_char_lens(self.src_file ) UpperCAmelCase__ : int = max_source_length UpperCAmelCase__ : List[str] = max_target_length assert min(self.src_lens ) > 0, f"found empty line in {self.src_file}" UpperCAmelCase__ : Dict = tokenizer UpperCAmelCase__ : str = prefix if n_obs is not None: UpperCAmelCase__ : int = self.src_lens[:n_obs] UpperCAmelCase__ : Any = src_lang UpperCAmelCase__ : Any = tgt_lang def __len__( self : Optional[Any] ): '''simple docstring''' return len(self.src_lens ) def __getitem__( self : Union[str, Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = index + 1 # linecache starts at 1 UpperCAmelCase__ : Tuple = self.prefix + linecache.getline(str(self.src_file ) ,A ).rstrip("""\n""" ) UpperCAmelCase__ : Dict = linecache.getline(str(self.tgt_file ) ,A ).rstrip("""\n""" ) assert source_line, f"empty source line for index {index}" assert tgt_line, f"empty tgt line for index {index}" # Need to add eos token manually for T5 if isinstance(self.tokenizer ,A ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right UpperCAmelCase__ : str = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,A ) else self.tokenizer ) UpperCAmelCase__ : Tuple = self.tokenizer.generator if isinstance(self.tokenizer ,A ) else self.tokenizer UpperCAmelCase__ : Tuple = encode_line(A ,A ,self.max_source_length ,"""right""" ) UpperCAmelCase__ : Dict = encode_line(A ,A ,self.max_target_length ,"""right""" ) UpperCAmelCase__ : Optional[Any] = source_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : List[str] = target_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : Union[str, Any] = source_inputs["""attention_mask"""].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def __lowercase ( A : int ): '''simple docstring''' return [len(A ) for x in Path(A ).open().readlines()] def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : int = torch.stack([x["""input_ids"""] for x in batch] ) UpperCAmelCase__ : Union[str, Any] = torch.stack([x["""attention_mask"""] for x in batch] ) UpperCAmelCase__ : Any = torch.stack([x["""decoder_input_ids"""] for x in batch] ) UpperCAmelCase__ : List[Any] = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : Any = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : str = trim_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = trim_batch(A ,A ,attention_mask=A ) UpperCAmelCase__ : List[str] = { """input_ids""": source_ids, """attention_mask""": source_mask, """decoder_input_ids""": y, } return batch __UpperCAmelCase = getLogger(__name__) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return list(itertools.chain.from_iterable(__UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = get_git_info() save_json(__UpperCamelCase , os.path.join(__UpperCamelCase , """git_log.json""" ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=4 , **__UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=__UpperCamelCase , **__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = git.Repo(search_parent_directories=__UpperCamelCase ) UpperCAmelCase__ : List[str] = { """repo_id""": str(__UpperCamelCase ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), """hostname""": str(socket.gethostname() ), } return repo_infos def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return list(map(__UpperCamelCase , __UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """wb""" ) as f: return pickle.dump(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' def remove_articles(__UpperCamelCase ): return re.sub(r"""\b(a|an|the)\b""" , """ """ , __UpperCamelCase ) def white_space_fix(__UpperCamelCase ): return " ".join(text.split() ) def remove_punc(__UpperCamelCase ): UpperCAmelCase__ : List[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCamelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCamelCase ) ) ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : Dict = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : int = Counter(__UpperCamelCase ) & Counter(__UpperCamelCase ) UpperCAmelCase__ : List[str] = sum(common.values() ) if num_same == 0: return 0 UpperCAmelCase__ : str = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Tuple = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return normalize_answer(__UpperCamelCase ) == normalize_answer(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' assert len(__UpperCamelCase ) == len(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 0 for hypo, pred in zip(__UpperCamelCase , __UpperCamelCase ): em += exact_match_score(__UpperCamelCase , __UpperCamelCase ) if len(__UpperCamelCase ) > 0: em /= len(__UpperCamelCase ) return {"em": em} def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return model_prefix.startswith("""rag""" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead UpperCAmelCase__ : str = """dropout_rate""" for p in extra_params: if getattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): if not hasattr(__UpperCamelCase , __UpperCamelCase ) and not hasattr(__UpperCamelCase , equivalent_param[p] ): logger.info("""config doesn't have a `{}` attribute""".format(__UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) continue UpperCAmelCase__ : Tuple = p if hasattr(__UpperCamelCase , __UpperCamelCase ) else equivalent_param[p] setattr(__UpperCamelCase , __UpperCamelCase , getattr(__UpperCamelCase , __UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) return hparams, config
65
1
"""simple docstring""" import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class __lowercase ( __lowerCamelCase ): snake_case_ = """""" snake_case_ = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) snake_case_ = None # compression type in fsspec. ex: "gzip" snake_case_ = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : Optional[int] ,A : str = "" ,A : Optional[str] = None ,A : Optional[dict] = None ,**A : Any ): '''simple docstring''' super().__init__(self ,**A ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode UpperCAmelCase__ : Optional[int] = fsspec.open( A ,mode="""rb""" ,protocol=A ,compression=self.compression ,client_kwargs={ """requote_redirect_url""": False, # see https://github.com/huggingface/datasets/pull/5459 """trust_env""": True, # Enable reading proxy env variables. **(target_options or {}).pop("""client_kwargs""" ,{} ), # To avoid issues if it was already passed. } ,**(target_options or {}) ,) UpperCAmelCase__ : int = os.path.basename(self.file.path.split("""::""" )[0] ) UpperCAmelCase__ : List[Any] = ( self.compressed_name[: self.compressed_name.rindex(""".""" )] if """.""" in self.compressed_name else self.compressed_name ) UpperCAmelCase__ : Optional[int] = None @classmethod def __lowercase ( cls : Optional[Any] ,A : Any ): '''simple docstring''' # compressed file paths are always relative to the archive root return super()._strip_protocol(A ).lstrip("""/""" ) def __lowercase ( self : str ): '''simple docstring''' if self.dir_cache is None: UpperCAmelCase__ : List[str] = {**self.file.fs.info(self.file.path ), """name""": self.uncompressed_name} UpperCAmelCase__ : List[str] = {f["""name"""]: f} def __lowercase ( self : Dict ,A : str ): '''simple docstring''' return self.file.open().read() def __lowercase ( self : Optional[int] ,A : str ,A : str = "rb" ,A : str=None ,A : List[Any]=True ,A : List[str]=None ,**A : Tuple ,): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self._strip_protocol(A ) if mode != "rb": raise ValueError(f"Tried to read with mode {mode} on file {self.file.path} opened with mode 'rb'" ) return self.file.open() class __lowercase ( __lowerCamelCase ): snake_case_ = """bz2""" snake_case_ = """bz2""" snake_case_ = """.bz2""" class __lowercase ( __lowerCamelCase ): snake_case_ = """gzip""" snake_case_ = """gzip""" snake_case_ = """.gz""" class __lowercase ( __lowerCamelCase ): snake_case_ = """lz4""" snake_case_ = """lz4""" snake_case_ = """.lz4""" class __lowercase ( __lowerCamelCase ): snake_case_ = """xz""" snake_case_ = """xz""" snake_case_ = """.xz""" class __lowercase ( __lowerCamelCase ): snake_case_ = """zstd""" snake_case_ = """zstd""" snake_case_ = """.zst""" def __init__( self : Union[str, Any] ,A : str ,A : str = "rb" ,A : Optional[str] = None ,A : Optional[dict] = None ,A : int = DEFAULT_BLOCK_SIZE ,**A : List[str] ,): '''simple docstring''' super().__init__( fo=A ,mode=A ,target_protocol=A ,target_options=A ,block_size=A ,**A ,) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 UpperCAmelCase__ : str = self.file.__enter__ class __lowercase : def __init__( self : Tuple ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = file_ def __enter__( self : List[Any] ): '''simple docstring''' self._file.__enter__() return self def __exit__( self : List[Any] ,*A : Union[str, Any] ,**A : Tuple ): '''simple docstring''' self._file.__exit__(*A ,**A ) def __iter__( self : Tuple ): '''simple docstring''' return iter(self._file ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' return next(self._file ) def __getattr__( self : Tuple ,A : Any ): '''simple docstring''' return getattr(self._file ,A ) def fixed_enter(*A : List[Any] ,**A : List[str] ): return WrappedFile(_enter(*A ,**A ) ) UpperCAmelCase__ : str = fixed_enter
65
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaControlnetPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return 32 @property def __lowercase ( self : int ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self : Any ): '''simple docstring''' return 100 @property def __lowercase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Tuple = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } UpperCAmelCase__ : int = UNetaDConditionModel(**A ) return model @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : str = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : str = self.dummy_unet UpperCAmelCase__ : List[Any] = self.dummy_movq UpperCAmelCase__ : List[Any] = DDIMScheduler( num_train_timesteps=1_000 ,beta_schedule="""linear""" ,beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,clip_sample=A ,set_alpha_to_one=A ,steps_offset=1 ,prediction_type="""epsilon""" ,thresholding=A ,) UpperCAmelCase__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __lowercase ( self : str ,A : Optional[Any] ,A : Any=0 ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( A ) # create hint UpperCAmelCase__ : int = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) if str(A ).startswith("""mps""" ): UpperCAmelCase__ : Optional[int] = torch.manual_seed(A ) else: UpperCAmelCase__ : Dict = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase__ : Dict = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = """cpu""" UpperCAmelCase__ : List[Any] = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = self.pipeline_class(**A ) UpperCAmelCase__ : Optional[int] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = pipe(**self.get_dummy_inputs(A ) ) UpperCAmelCase__ : Tuple = output.images UpperCAmelCase__ : Dict = pipe( **self.get_dummy_inputs(A ) ,return_dict=A ,)[0] UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1] UpperCAmelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Optional[int] = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : Union[str, Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) UpperCAmelCase__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) UpperCAmelCase__ : int = torch.from_numpy(np.array(A ) ).float() / 2_5_5.0 UpperCAmelCase__ : Union[str, Any] = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) UpperCAmelCase__ : List[str] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" ,torch_dtype=torch.floataa ) pipe_prior.to(A ) UpperCAmelCase__ : List[Any] = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" ,torch_dtype=torch.floataa ) UpperCAmelCase__ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[Any] = """A robot, 4k photo""" UpperCAmelCase__ : List[Any] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = pipe_prior( A ,generator=A ,num_inference_steps=5 ,negative_prompt="""""" ,).to_tuple() UpperCAmelCase__ : List[str] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ : int = pipeline( image_embeds=A ,negative_image_embeds=A ,hint=A ,generator=A ,num_inference_steps=100 ,output_type="""np""" ,) UpperCAmelCase__ : Any = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A ,A )
65
1
"""simple docstring""" import math def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(__UpperCamelCase ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("""This should never happen""" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. __UpperCAmelCase = 'Enter the base and the power separated by a comma: ' __UpperCAmelCase, __UpperCAmelCase = map(int, input(prompt).split(',')) __UpperCAmelCase, __UpperCAmelCase = map(int, input(prompt).split(',')) # We find the log of each number, using the function res(), which takes two # arguments. __UpperCAmelCase = res(xa, ya) __UpperCAmelCase = res(xa, ya) # We check for the largest number if resa > resa: print('Largest number is', xa, '^', ya) elif resa > resa: print('Largest number is', xa, '^', ya) else: print('Both are equal')
65
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = """vision-encoder-decoder""" snake_case_ = True def __init__( self : List[Any] ,**A : Union[str, Any] ): '''simple docstring''' super().__init__(**A ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) UpperCAmelCase__ : int = kwargs.pop("""encoder""" ) UpperCAmelCase__ : int = encoder_config.pop("""model_type""" ) UpperCAmelCase__ : str = kwargs.pop("""decoder""" ) UpperCAmelCase__ : Dict = decoder_config.pop("""model_type""" ) UpperCAmelCase__ : List[Any] = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Any = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Union[str, Any] = True @classmethod def __lowercase ( cls : List[Any] ,A : PretrainedConfig ,A : PretrainedConfig ,**A : Tuple ): '''simple docstring''' logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : List[Any] = True return cls(encoder=encoder_config.to_dict() ,decoder=decoder_config.to_dict() ,**A ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Dict = self.encoder.to_dict() UpperCAmelCase__ : Any = self.decoder.to_dict() UpperCAmelCase__ : Dict = self.__class__.model_type return output class __lowercase ( __lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' return 1e-4 @property def __lowercase ( self : List[Any] ): '''simple docstring''' return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : Optional[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : List[str] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def __lowercase ( self : Dict ,A : "PreTrainedTokenizerBase" ,A : int = -1 ,A : int = -1 ,A : bool = False ,A : Optional["TensorType"] = None ,): '''simple docstring''' import torch UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : List[Any] = super().generate_dummy_inputs( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) UpperCAmelCase__ , UpperCAmelCase__ : int = dummy_input["""input_ids"""].shape UpperCAmelCase__ : int = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCAmelCase__ : Tuple = dummy_input.pop("""input_ids""" ) UpperCAmelCase__ : Optional[int] = dummy_input.pop("""attention_mask""" ) UpperCAmelCase__ : Dict = torch.zeros(A ) return common_inputs class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : Any ,A : PretrainedConfig ): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(A ) def __lowercase ( self : Dict ,A : PretrainedConfig ,A : PretrainedConfig ,A : str = "default" ): '''simple docstring''' UpperCAmelCase__ : List[str] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(A ,A )
65
1
"""simple docstring""" import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput __UpperCAmelCase = 'scheduler_config.json' class __lowercase ( __lowerCamelCase ): snake_case_ = 1 snake_case_ = 2 snake_case_ = 3 snake_case_ = 4 snake_case_ = 5 @dataclass class __lowercase ( __lowerCamelCase ): snake_case_ = 42 class __lowercase : snake_case_ = SCHEDULER_CONFIG_NAME snake_case_ = ["""dtype"""] snake_case_ = [] snake_case_ = True @classmethod def __lowercase ( cls : List[str] ,A : Dict[str, Any] = None ,A : Optional[str] = None ,A : List[Any]=False ,**A : Union[str, Any] ,): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = cls.load_config( pretrained_model_name_or_path=A ,subfolder=A ,return_unused_kwargs=A ,**A ,) UpperCAmelCase__ , UpperCAmelCase__ : int = cls.from_config(A ,return_unused_kwargs=A ,**A ) if hasattr(A ,"""create_state""" ) and getattr(A ,"""has_state""" ,A ): UpperCAmelCase__ : Any = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def __lowercase ( self : int ,A : Union[str, os.PathLike] ,A : bool = False ,**A : Dict ): '''simple docstring''' self.save_config(save_directory=A ,push_to_hub=A ,**A ) @property def __lowercase ( self : Optional[Any] ): '''simple docstring''' return self._get_compatibles() @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' UpperCAmelCase__ : int = list(set([cls.__name__] + cls._compatibles ) ) UpperCAmelCase__ : Tuple = importlib.import_module(__name__.split(""".""" )[0] ) UpperCAmelCase__ : Tuple = [ getattr(A ,A ) for c in compatible_classes_str if hasattr(A ,A ) ] return compatible_classes def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' assert len(__UpperCamelCase ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__UpperCamelCase ) - x.ndim) ) , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase=0.999 , __UpperCamelCase=jnp.floataa ): '''simple docstring''' def alpha_bar(__UpperCamelCase ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 UpperCAmelCase__ : Optional[Any] = [] for i in range(__UpperCamelCase ): UpperCAmelCase__ : Dict = i / num_diffusion_timesteps UpperCAmelCase__ : str = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__UpperCamelCase ) / alpha_bar(__UpperCamelCase ) , __UpperCamelCase ) ) return jnp.array(__UpperCamelCase , dtype=__UpperCamelCase ) @flax.struct.dataclass class __lowercase : snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 @classmethod def __lowercase ( cls : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = scheduler.config if config.trained_betas is not None: UpperCAmelCase__ : Any = jnp.asarray(config.trained_betas ,dtype=scheduler.dtype ) elif config.beta_schedule == "linear": UpperCAmelCase__ : Optional[Any] = jnp.linspace(config.beta_start ,config.beta_end ,config.num_train_timesteps ,dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. UpperCAmelCase__ : List[str] = ( jnp.linspace( config.beta_start**0.5 ,config.beta_end**0.5 ,config.num_train_timesteps ,dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule UpperCAmelCase__ : Dict = betas_for_alpha_bar(config.num_train_timesteps ,dtype=scheduler.dtype ) else: raise NotImplementedError( f"beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}" ) UpperCAmelCase__ : Optional[Any] = 1.0 - betas UpperCAmelCase__ : List[str] = jnp.cumprod(A ,axis=0 ) return cls( alphas=A ,betas=A ,alphas_cumprod=A ,) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = state.alphas_cumprod UpperCAmelCase__ : Optional[int] = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase__ : Dict = sqrt_alpha_prod.flatten() UpperCAmelCase__ : Union[str, Any] = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) UpperCAmelCase__ : Any = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase__ : Any = sqrt_one_minus_alpha_prod.flatten() UpperCAmelCase__ : int = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) UpperCAmelCase__ : List[str] = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
65
"""simple docstring""" import requests def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = {"""Content-Type""": """application/json"""} UpperCAmelCase__ : Optional[Any] = requests.post(__UpperCamelCase , json={"""text""": message_body} , headers=__UpperCamelCase ) if response.status_code != 200: UpperCAmelCase__ : Any = ( """Request to slack returned an error """ F"{response.status_code}, the response is:\n{response.text}" ) raise ValueError(__UpperCamelCase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
65
1
"""simple docstring""" import argparse from collections import defaultdict import yaml __UpperCAmelCase = 'docs/source/en/_toctree.yml' def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = defaultdict(__UpperCamelCase ) UpperCAmelCase__ : Any = [] UpperCAmelCase__ : Tuple = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({"""local""": doc["""local"""], """title""": doc["""title"""]} ) else: new_doc_list.append(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = new_doc_list UpperCAmelCase__ : Tuple = [key for key, value in counts.items() if value > 1] UpperCAmelCase__ : str = [] for duplicate_key in duplicates: UpperCAmelCase__ : Tuple = list({doc["""title"""] for doc in doc_list if doc["""local"""] == duplicate_key} ) if len(__UpperCamelCase ) > 1: raise ValueError( F"{duplicate_key} is present several times in the documentation table of content at " """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if """local""" not in counts or counts[doc["""local"""]] == 1] ) UpperCAmelCase__ : List[str] = sorted(__UpperCamelCase , key=lambda __UpperCamelCase : s["title"].lower() ) # "overview" gets special treatment and is always first if len(__UpperCamelCase ) > 1: raise ValueError("""{doc_list} has two 'overview' docs which is not allowed.""" ) overview_doc.extend(__UpperCamelCase ) # Sort return overview_doc def lowerCAmelCase ( __UpperCamelCase=False ): '''simple docstring''' with open(__UpperCamelCase , encoding="""utf-8""" ) as f: UpperCAmelCase__ : Tuple = yaml.safe_load(f.read() ) # Get to the API doc UpperCAmelCase__ : Tuple = 0 while content[api_idx]["title"] != "API": api_idx += 1 UpperCAmelCase__ : Dict = content[api_idx]["""sections"""] # Then to the model doc UpperCAmelCase__ : Any = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 UpperCAmelCase__ : Union[str, Any] = api_doc[scheduler_idx]["""sections"""] UpperCAmelCase__ : Dict = clean_doc_toc(__UpperCamelCase ) UpperCAmelCase__ : Dict = False if new_scheduler_doc != scheduler_doc: UpperCAmelCase__ : Optional[Any] = True if overwrite: UpperCAmelCase__ : int = new_scheduler_doc if diff: if overwrite: UpperCAmelCase__ : Any = api_doc with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__UpperCamelCase , allow_unicode=__UpperCamelCase ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) def lowerCAmelCase ( __UpperCamelCase=False ): '''simple docstring''' with open(__UpperCamelCase , encoding="""utf-8""" ) as f: UpperCAmelCase__ : List[str] = yaml.safe_load(f.read() ) # Get to the API doc UpperCAmelCase__ : Optional[Any] = 0 while content[api_idx]["title"] != "API": api_idx += 1 UpperCAmelCase__ : Tuple = content[api_idx]["""sections"""] # Then to the model doc UpperCAmelCase__ : List[str] = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 UpperCAmelCase__ : str = False UpperCAmelCase__ : List[Any] = api_doc[pipeline_idx]["""sections"""] UpperCAmelCase__ : str = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: UpperCAmelCase__ : Optional[Any] = pipeline_doc["""section"""] UpperCAmelCase__ : Union[str, Any] = clean_doc_toc(__UpperCamelCase ) if overwrite: UpperCAmelCase__ : str = new_sub_pipeline_doc new_pipeline_docs.append(__UpperCamelCase ) # sort overall pipeline doc UpperCAmelCase__ : List[str] = clean_doc_toc(__UpperCamelCase ) if new_pipeline_docs != pipeline_docs: UpperCAmelCase__ : Any = True if overwrite: UpperCAmelCase__ : List[str] = new_pipeline_docs if diff: if overwrite: UpperCAmelCase__ : Optional[int] = api_doc with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__UpperCamelCase , allow_unicode=__UpperCamelCase ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') __UpperCAmelCase = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
65
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = CTRLTokenizer snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ : List[Any] = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase__ : Optional[int] = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : List[Any] = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase__ : int = {"""unk_token""": """<unk>"""} UpperCAmelCase__ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(A ) ) def __lowercase ( self : int ,**A : Dict ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """adapt react readapt apt""" UpperCAmelCase__ : Any = """adapt react readapt apt""" return input_text, output_text def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = CTRLTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) UpperCAmelCase__ : Tuple = """adapt react readapt apt""" UpperCAmelCase__ : Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase__ : Dict = tokenizer.tokenize(A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Any = tokens + [tokenizer.unk_token] UpperCAmelCase__ : Dict = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,A )
65
1
"""simple docstring""" from __future__ import annotations def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return [ord(__UpperCamelCase ) - 96 for elem in plain] def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return "".join(chr(elem + 96 ) for elem in encoded ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = encode(input("""-> """ ).strip().lower() ) print("""Encoded: """ , __UpperCamelCase ) print("""Decoded:""" , decode(__UpperCamelCase ) ) if __name__ == "__main__": main()
65
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCAmelCase = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfig', 'BridgeTowerTextConfig', 'BridgeTowerVisionConfig', ], 'processing_bridgetower': ['BridgeTowerProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['BridgeTowerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST', 'BridgeTowerForContrastiveLearning', 'BridgeTowerForImageAndTextRetrieval', 'BridgeTowerForMaskedLM', 'BridgeTowerModel', 'BridgeTowerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
65
1
"""simple docstring""" import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) __UpperCAmelCase = logging.getLogger(__name__) __UpperCAmelCase = 'Hello world! cécé herlolip' __UpperCAmelCase = namedtuple( 'BertAbsConfig', [ 'temp_dir', 'large', 'use_bert_emb', 'finetune_bert', 'encoder', 'share_emb', 'max_pos', 'enc_layers', 'enc_hidden_size', 'enc_heads', 'enc_ff_size', 'enc_dropout', 'dec_layers', 'dec_hidden_size', 'dec_heads', 'dec_ff_size', 'dec_dropout', ], ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = BertAbsConfig( temp_dir=""".""" , finetune_bert=__UpperCamelCase , large=__UpperCamelCase , share_emb=__UpperCamelCase , use_bert_emb=__UpperCamelCase , encoder="""bert""" , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) UpperCAmelCase__ : Dict = torch.load(__UpperCamelCase , lambda __UpperCamelCase , __UpperCamelCase : storage ) UpperCAmelCase__ : int = AbsSummarizer(__UpperCamelCase , torch.device("""cpu""" ) , __UpperCamelCase ) original.eval() UpperCAmelCase__ : int = BertAbsSummarizer(__UpperCamelCase , torch.device("""cpu""" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("""convert the model""" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("""Make sure that the models' outputs are identical""" ) UpperCAmelCase__ : List[Any] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) # prepare the model inputs UpperCAmelCase__ : List[Any] = tokenizer.encode("""This is sample éàalj'-.""" ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(__UpperCamelCase )) ) UpperCAmelCase__ : List[Any] = torch.tensor(__UpperCamelCase ).unsqueeze(0 ) UpperCAmelCase__ : Any = tokenizer.encode("""This is sample 3 éàalj'-.""" ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(__UpperCamelCase )) ) UpperCAmelCase__ : List[str] = torch.tensor(__UpperCamelCase ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass UpperCAmelCase__ : List[Any] = encoder_input_ids UpperCAmelCase__ : Optional[Any] = decoder_input_ids UpperCAmelCase__ : str = None UpperCAmelCase__ : Optional[int] = None UpperCAmelCase__ : List[str] = None UpperCAmelCase__ : Tuple = None UpperCAmelCase__ : int = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical UpperCAmelCase__ : Tuple = original(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )[0] UpperCAmelCase__ : Dict = original.generator(__UpperCamelCase ) UpperCAmelCase__ : List[Any] = new_model( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )[0] UpperCAmelCase__ : Any = new_model.generator(__UpperCamelCase ) UpperCAmelCase__ : int = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("""Maximum absolute difference beween weights: {:.2f}""".format(__UpperCamelCase ) ) UpperCAmelCase__ : Union[str, Any] = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("""Maximum absolute difference beween weights: {:.2f}""".format(__UpperCamelCase ) ) UpperCAmelCase__ : Union[str, Any] = torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1e-3 ) if are_identical: logging.info("""all weights are equal up to 1e-3""" ) else: raise ValueError("""the weights are different. The new model is likely different from the original one.""" ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("""saving the model's state dictionary""" ) torch.save( new_model.state_dict() , """./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( '--bertabs_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.', ) __UpperCAmelCase = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
65
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """is_longer"""] def __init__( self : str ,A : Union[str, Any]=64 ,A : Tuple=48_000 ,A : Dict=480 ,A : List[str]=10 ,A : str=1_024 ,A : Any=0.0 ,A : Optional[int]=False ,A : float = 0 ,A : float = 14_000 ,A : int = None ,A : str = "fusion" ,A : str = "repeatpad" ,**A : List[Any] ,): '''simple docstring''' super().__init__( feature_size=A ,sampling_rate=A ,padding_value=A ,return_attention_mask=A ,**A ,) UpperCAmelCase__ : List[Any] = top_db UpperCAmelCase__ : Union[str, Any] = truncation UpperCAmelCase__ : Optional[int] = padding UpperCAmelCase__ : List[Any] = fft_window_size UpperCAmelCase__ : Optional[Any] = (fft_window_size >> 1) + 1 UpperCAmelCase__ : Any = hop_length UpperCAmelCase__ : List[str] = max_length_s UpperCAmelCase__ : List[Any] = max_length_s * sampling_rate UpperCAmelCase__ : List[Any] = sampling_rate UpperCAmelCase__ : Optional[int] = frequency_min UpperCAmelCase__ : Tuple = frequency_max UpperCAmelCase__ : List[str] = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm=A ,mel_scale="""htk""" ,) UpperCAmelCase__ : str = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm="""slaney""" ,mel_scale="""slaney""" ,) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def __lowercase ( self : List[str] ,A : np.array ,A : Optional[np.array] = None ): '''simple docstring''' UpperCAmelCase__ : Dict = spectrogram( A ,window_function(self.fft_window_size ,"""hann""" ) ,frame_length=self.fft_window_size ,hop_length=self.hop_length ,power=2.0 ,mel_filters=A ,log_mel="""dB""" ,) return log_mel_spectrogram.T def __lowercase ( self : Optional[Any] ,A : Union[str, Any] ,A : int ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = np.array_split(list(range(0 ,total_frames - chunk_frames + 1 ) ) ,3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : List[str] = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : int = [0] # randomly choose index for each part UpperCAmelCase__ : Tuple = np.random.choice(ranges[0] ) UpperCAmelCase__ : Tuple = np.random.choice(ranges[1] ) UpperCAmelCase__ : str = np.random.choice(ranges[2] ) UpperCAmelCase__ : List[str] = mel[idx_front : idx_front + chunk_frames, :] UpperCAmelCase__ : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] UpperCAmelCase__ : Dict = mel[idx_back : idx_back + chunk_frames, :] UpperCAmelCase__ : Optional[Any] = torch.tensor(mel[None, None, :] ) UpperCAmelCase__ : int = torch.nn.functional.interpolate( A ,size=[chunk_frames, 64] ,mode="""bilinear""" ,align_corners=A ) UpperCAmelCase__ : Dict = mel_shrink[0][0].numpy() UpperCAmelCase__ : Dict = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] ,axis=0 ) return mel_fusion def __lowercase ( self : Any ,A : np.array ,A : Optional[int] ,A : Any ,A : Tuple ): '''simple docstring''' if waveform.shape[0] > max_length: if truncation == "rand_trunc": UpperCAmelCase__ : int = True # random crop to max_length (for compatibility) -> this should be handled by self.pad UpperCAmelCase__ : str = len(A ) - max_length UpperCAmelCase__ : Optional[Any] = np.random.randint(0 ,overflow + 1 ) UpperCAmelCase__ : Optional[int] = waveform[idx : idx + max_length] UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] elif truncation == "fusion": UpperCAmelCase__ : Tuple = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : Optional[int] = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed UpperCAmelCase__ : int = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. UpperCAmelCase__ : List[Any] = np.stack([mel, mel, mel, mel] ,axis=0 ) UpperCAmelCase__ : Any = False else: UpperCAmelCase__ : Union[str, Any] = self._random_mel_fusion(A ,A ,A ) UpperCAmelCase__ : List[str] = True else: raise NotImplementedError(f"data_truncating {truncation} not implemented" ) else: UpperCAmelCase__ : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": UpperCAmelCase__ : str = int(max_length / len(A ) ) UpperCAmelCase__ : int = np.stack(np.tile(A ,n_repeat + 1 ) )[:max_length] if padding == "repeatpad": UpperCAmelCase__ : List[Any] = int(max_length / len(A ) ) UpperCAmelCase__ : str = np.stack(np.tile(A ,A ) ) UpperCAmelCase__ : Optional[Any] = np.pad(A ,(0, max_length - waveform.shape[0]) ,mode="""constant""" ,constant_values=0 ) if truncation == "fusion": UpperCAmelCase__ : int = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] ,axis=0 ) else: UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self : str ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : str = None ,A : Optional[str] = None ,A : Optional[int] = None ,A : Optional[int] = None ,A : Optional[Union[str, TensorType]] = None ,**A : List[str] ,): '''simple docstring''' UpperCAmelCase__ : Optional[int] = truncation if truncation is not None else self.truncation UpperCAmelCase__ : Dict = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" f" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" f" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[int] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : List[str] = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Any = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [np.asarray(A )] # convert to mel spectrogram, truncate and pad if needed. UpperCAmelCase__ : Tuple = [ self._get_input_mel(A ,max_length if max_length else self.nb_max_samples ,A ,A ) for waveform in raw_speech ] UpperCAmelCase__ : Optional[int] = [] UpperCAmelCase__ : Tuple = [] for mel, longer in padded_inputs: input_mel.append(A ) is_longer.append(A ) if truncation == "fusion" and sum(A ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer UpperCAmelCase__ : List[str] = np.random.randint(0 ,len(A ) ) UpperCAmelCase__ : int = True if isinstance(input_mel[0] ,A ): UpperCAmelCase__ : Tuple = [np.asarray(A ,dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool UpperCAmelCase__ : List[str] = [[longer] for longer in is_longer] UpperCAmelCase__ : List[Any] = {"""input_features""": input_mel, """is_longer""": is_longer} UpperCAmelCase__ : str = BatchFeature(A ) if return_tensors is not None: UpperCAmelCase__ : int = input_features.convert_to_tensors(A ) return input_features
65
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'facebook/data2vec-text-base': 'https://huggingface.co/data2vec/resolve/main/config.json', } class __lowercase ( __lowerCamelCase ): snake_case_ = """data2vec-text""" def __init__( self : str ,A : Union[str, Any]=30_522 ,A : Union[str, Any]=768 ,A : Tuple=12 ,A : Dict=12 ,A : Dict=3_072 ,A : Any="gelu" ,A : Union[str, Any]=0.1 ,A : int=0.1 ,A : int=512 ,A : Union[str, Any]=2 ,A : Tuple=0.0_2 ,A : Dict=1e-12 ,A : Optional[int]=1 ,A : Tuple=0 ,A : Optional[int]=2 ,A : Tuple="absolute" ,A : Any=True ,A : Optional[int]=None ,**A : Union[str, Any] ,): '''simple docstring''' super().__init__(pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,**A ) UpperCAmelCase__ : Union[str, Any] = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Dict = num_hidden_layers UpperCAmelCase__ : Optional[int] = num_attention_heads UpperCAmelCase__ : Tuple = hidden_act UpperCAmelCase__ : List[Any] = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : Optional[int] = attention_probs_dropout_prob UpperCAmelCase__ : Any = max_position_embeddings UpperCAmelCase__ : Optional[int] = type_vocab_size UpperCAmelCase__ : Dict = initializer_range UpperCAmelCase__ : Union[str, Any] = layer_norm_eps UpperCAmelCase__ : List[Any] = position_embedding_type UpperCAmelCase__ : Union[str, Any] = use_cache UpperCAmelCase__ : Tuple = classifier_dropout class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : int ): '''simple docstring''' if self.task == "multiple-choice": UpperCAmelCase__ : Dict = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase__ : Union[str, Any] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
65
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Union[str, Any] ,A : Dict=7 ,A : Optional[int]=3 ,A : List[str]=18 ,A : Union[str, Any]=30 ,A : Tuple=400 ,A : Dict=True ,A : List[str]=None ,A : str=True ,A : Optional[Any]=False ,A : Optional[Any]=True ,A : List[str]=True ,A : Optional[int]=[0.5, 0.5, 0.5] ,A : List[str]=[0.5, 0.5, 0.5] ,): '''simple docstring''' UpperCAmelCase__ : str = parent UpperCAmelCase__ : List[str] = batch_size UpperCAmelCase__ : List[str] = num_channels UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : Optional[int] = max_resolution UpperCAmelCase__ : str = do_resize UpperCAmelCase__ : Tuple = size if size is not None else {"""height""": 18, """width""": 20} UpperCAmelCase__ : List[str] = do_thumbnail UpperCAmelCase__ : Optional[int] = do_align_axis UpperCAmelCase__ : Union[str, Any] = do_pad UpperCAmelCase__ : Tuple = do_normalize UpperCAmelCase__ : Optional[Any] = image_mean UpperCAmelCase__ : List[Any] = image_std def __lowercase ( self : Optional[int] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = DonutImageProcessor if is_vision_available() else None def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = DonutImageProcessingTester(self ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_thumbnail""" ) ) self.assertTrue(hasattr(A ,"""do_align_long_axis""" ) ) self.assertTrue(hasattr(A ,"""do_pad""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 18, """width""": 20} ) UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=(42, 84) ) self.assertEqual(image_processor.size ,{"""height""": 84, """width""": 42} ) def __lowercase ( self : Dict ): '''simple docstring''' pass @is_flaky() def __lowercase ( self : int ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase__ : int = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Tuple = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : List[str] ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase__ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Optional[int] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : Any ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : List[Any] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,)
65
1
"""simple docstring""" import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) __UpperCAmelCase = { 'iou_prediction_head.layers.0': 'iou_prediction_head.proj_in', 'iou_prediction_head.layers.1': 'iou_prediction_head.layers.0', 'iou_prediction_head.layers.2': 'iou_prediction_head.proj_out', 'mask_decoder.output_upscaling.0': 'mask_decoder.upscale_conv1', 'mask_decoder.output_upscaling.1': 'mask_decoder.upscale_layer_norm', 'mask_decoder.output_upscaling.3': 'mask_decoder.upscale_conv2', 'mask_downscaling.0': 'mask_embed.conv1', 'mask_downscaling.1': 'mask_embed.layer_norm1', 'mask_downscaling.3': 'mask_embed.conv2', 'mask_downscaling.4': 'mask_embed.layer_norm2', 'mask_downscaling.6': 'mask_embed.conv3', 'point_embeddings': 'point_embed', 'pe_layer.positional_encoding_gaussian_matrix': 'shared_embedding.positional_embedding', 'image_encoder': 'vision_encoder', 'neck.0': 'neck.conv1', 'neck.1': 'neck.layer_norm1', 'neck.2': 'neck.conv2', 'neck.3': 'neck.layer_norm2', 'patch_embed.proj': 'patch_embed.projection', '.norm': '.layer_norm', 'blocks': 'layers', } def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {} state_dict.pop("""pixel_mean""" , __UpperCamelCase ) state_dict.pop("""pixel_std""" , __UpperCamelCase ) UpperCAmelCase__ : int = r""".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*""" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase__ : List[str] = key.replace(__UpperCamelCase , __UpperCamelCase ) if re.match(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Tuple = int(re.match(__UpperCamelCase , __UpperCamelCase ).group(2 ) ) if layer_nb == 0: UpperCAmelCase__ : Optional[Any] = key.replace("""layers.0""" , """proj_in""" ) elif layer_nb == 1: UpperCAmelCase__ : Any = key.replace("""layers.1""" , """layers.0""" ) elif layer_nb == 2: UpperCAmelCase__ : Any = key.replace("""layers.2""" , """proj_out""" ) UpperCAmelCase__ : Union[str, Any] = value UpperCAmelCase__ : Dict = model_state_dict[ """prompt_encoder.shared_embedding.positional_embedding""" ] return model_state_dict def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase="ybelkada/segment-anything" ): '''simple docstring''' UpperCAmelCase__ : str = hf_hub_download(__UpperCamelCase , F"checkpoints/{model_name}.pth" ) if "sam_vit_b" in model_name: UpperCAmelCase__ : Any = SamConfig() elif "sam_vit_l" in model_name: UpperCAmelCase__ : Union[str, Any] = SamVisionConfig( hidden_size=1024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) UpperCAmelCase__ : str = SamConfig( vision_config=__UpperCamelCase , ) elif "sam_vit_h" in model_name: UpperCAmelCase__ : List[Any] = SamVisionConfig( hidden_size=1280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) UpperCAmelCase__ : Optional[int] = SamConfig( vision_config=__UpperCamelCase , ) UpperCAmelCase__ : List[str] = torch.load(__UpperCamelCase , map_location="""cpu""" ) UpperCAmelCase__ : Tuple = replace_keys(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = SamImageProcessor() UpperCAmelCase__ : Dict = SamProcessor(image_processor=__UpperCamelCase ) UpperCAmelCase__ : int = SamModel(__UpperCamelCase ) hf_model.load_state_dict(__UpperCamelCase ) UpperCAmelCase__ : List[Any] = hf_model.to("""cuda""" ) UpperCAmelCase__ : Optional[Any] = """https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png""" UpperCAmelCase__ : Dict = Image.open(requests.get(__UpperCamelCase , stream=__UpperCamelCase ).raw ).convert("""RGB""" ) UpperCAmelCase__ : Tuple = [[[400, 650]]] UpperCAmelCase__ : Tuple = [[1]] UpperCAmelCase__ : int = processor(images=np.array(__UpperCamelCase ) , return_tensors="""pt""" ).to("""cuda""" ) with torch.no_grad(): UpperCAmelCase__ : Optional[int] = hf_model(**__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_8902_5115_9668 UpperCAmelCase__ : Union[str, Any] = processor( images=np.array(__UpperCamelCase ) , input_points=__UpperCamelCase , input_labels=__UpperCamelCase , return_tensors="""pt""" ).to("""cuda""" ) with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = hf_model(**__UpperCamelCase ) UpperCAmelCase__ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712_6030_9219_3604 UpperCAmelCase__ : str = ((75, 275, 1725, 850),) UpperCAmelCase__ : List[Any] = processor(images=np.array(__UpperCamelCase ) , input_boxes=__UpperCamelCase , return_tensors="""pt""" ).to("""cuda""" ) with torch.no_grad(): UpperCAmelCase__ : Optional[int] = hf_model(**__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686_0156_0592_6514 # Test with 2 points and 1 image. UpperCAmelCase__ : Any = [[[400, 650], [800, 650]]] UpperCAmelCase__ : int = [[1, 1]] UpperCAmelCase__ : str = processor( images=np.array(__UpperCamelCase ) , input_points=__UpperCamelCase , input_labels=__UpperCamelCase , return_tensors="""pt""" ).to("""cuda""" ) with torch.no_grad(): UpperCAmelCase__ : Tuple = hf_model(**__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936_0477_9243_4692 if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() __UpperCAmelCase = ['sam_vit_b_01ec64', 'sam_vit_h_4b8939', 'sam_vit_l_0b3195'] parser.add_argument( '--model_name', default='sam_vit_h_4b8939', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) parser.add_argument( '--model_hub_id', default='ybelkada/segment-anything', choices=choices, type=str, help='Path to hf config.json of model to convert', ) __UpperCAmelCase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
65
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class __lowercase ( __lowerCamelCase ): snake_case_ = """open-llama""" def __init__( self : Dict ,A : str=100_000 ,A : str=4_096 ,A : Optional[Any]=11_008 ,A : Tuple=32 ,A : str=32 ,A : Optional[int]="silu" ,A : List[Any]=2_048 ,A : str=0.0_2 ,A : Optional[int]=1e-6 ,A : int=True ,A : Tuple=0 ,A : str=1 ,A : Any=2 ,A : Optional[Any]=False ,A : int=True ,A : Any=0.1 ,A : Optional[Any]=0.1 ,A : Optional[Any]=True ,A : Union[str, Any]=True ,A : Tuple=None ,**A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : str = vocab_size UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : Tuple = intermediate_size UpperCAmelCase__ : Optional[int] = num_hidden_layers UpperCAmelCase__ : Any = num_attention_heads UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = initializer_range UpperCAmelCase__ : Optional[int] = rms_norm_eps UpperCAmelCase__ : Any = use_cache UpperCAmelCase__ : Optional[Any] = kwargs.pop( """use_memorry_efficient_attention""" ,A ) UpperCAmelCase__ : Any = hidden_dropout_prob UpperCAmelCase__ : str = attention_dropout_prob UpperCAmelCase__ : Optional[int] = use_stable_embedding UpperCAmelCase__ : Tuple = shared_input_output_embedding UpperCAmelCase__ : Tuple = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,tie_word_embeddings=A ,**A ,) def __lowercase ( self : Optional[Any] ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ f"got {self.rope_scaling}" ) UpperCAmelCase__ : List[Any] = self.rope_scaling.get("""type""" ,A ) UpperCAmelCase__ : int = self.rope_scaling.get("""factor""" ,A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(A ,A ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
65
1
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase ( __UpperCamelCase = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCAmelCase__ : Dict = F"{olid} is not a valid Open Library olid" raise ValueError(__UpperCamelCase ) return requests.get(F"https://openlibrary.org/{new_olid}.json" ).json() def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } UpperCAmelCase__ : Dict = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase__ : str = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] UpperCAmelCase__ : Dict = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Dict = """, """.join(__UpperCamelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __UpperCAmelCase = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: __UpperCAmelCase = summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('\n'.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
65
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
65
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """attention_mask"""] def __init__( self : Any ,A : str=80 ,A : Optional[int]=16_000 ,A : int=0.0 ,A : str=10 ,A : Any=25 ,A : str="hamming_window" ,A : int=3_2_7_6_8.0 ,A : List[str]=0.9_7 ,A : Optional[int]=1.0 ,A : Optional[Any]=True ,A : Tuple=True ,A : Any=False ,**A : int ,): '''simple docstring''' super().__init__(feature_size=A ,sampling_rate=A ,padding_value=A ,**A ) UpperCAmelCase__ : str = feature_size UpperCAmelCase__ : int = sampling_rate UpperCAmelCase__ : int = padding_value UpperCAmelCase__ : Dict = hop_length UpperCAmelCase__ : int = win_length UpperCAmelCase__ : Dict = frame_signal_scale UpperCAmelCase__ : Dict = preemphasis_coeff UpperCAmelCase__ : str = mel_floor UpperCAmelCase__ : Any = normalize_means UpperCAmelCase__ : str = normalize_vars UpperCAmelCase__ : int = win_function UpperCAmelCase__ : List[Any] = return_attention_mask UpperCAmelCase__ : str = win_length * sampling_rate // 1_000 UpperCAmelCase__ : List[Any] = hop_length * sampling_rate // 1_000 UpperCAmelCase__ : int = optimal_fft_length(self.sample_size ) UpperCAmelCase__ : List[Any] = (self.n_fft // 2) + 1 def __lowercase ( self : Union[str, Any] ,A : np.array ): '''simple docstring''' if self.win_function == "hamming_window": UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=A ) else: UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ) UpperCAmelCase__ : Union[str, Any] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) UpperCAmelCase__ : Optional[Any] = spectrogram( one_waveform * self.frame_signal_scale ,window=A ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=A ,preemphasis=self.preemphasis_coeff ,mel_filters=A ,mel_floor=self.mel_floor ,log_mel="""log""" ,) return msfc_features.T def __lowercase ( self : str ,A : Any ,A : Optional[int] ,A : str ): '''simple docstring''' # make sure we normalize float32 arrays if self.normalize_means: UpperCAmelCase__ : Optional[Any] = x[:input_length].mean(axis=0 ) UpperCAmelCase__ : Any = np.subtract(A ,A ) if self.normalize_vars: UpperCAmelCase__ : str = x[:input_length].std(axis=0 ) UpperCAmelCase__ : Optional[int] = np.divide(A ,A ) if input_length < x.shape[0]: UpperCAmelCase__ : int = padding_value # make sure array is in float32 UpperCAmelCase__ : str = x.astype(np.floataa ) return x def __lowercase ( self : Union[str, Any] ,A : List[np.ndarray] ,A : Optional[np.ndarray] = None ): '''simple docstring''' UpperCAmelCase__ : Any = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(A ,A ,self.padding_value ) for x, n in zip(A ,A )] def __call__( self : Union[str, Any] ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : Union[bool, str, PaddingStrategy] = False ,A : Optional[int] = None ,A : bool = False ,A : Optional[int] = None ,A : Optional[bool] = None ,A : Optional[Union[str, TensorType]] = None ,A : Optional[int] = None ,**A : Tuple ,): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" f" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" f" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[Any] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : Any = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : List[str] = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Union[str, Any] = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [raw_speech] # extract fbank features UpperCAmelCase__ : Tuple = [self._extract_mfsc_features(A ) for one_waveform in raw_speech] # convert into correct format for padding UpperCAmelCase__ : str = BatchFeature({"""input_features""": features} ) UpperCAmelCase__ : Optional[Any] = self.pad( A ,padding=A ,max_length=A ,truncation=A ,pad_to_multiple_of=A ,return_attention_mask=A ,**A ,) # make sure list is in array format UpperCAmelCase__ : Tuple = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] ,A ): UpperCAmelCase__ : Union[str, Any] = [np.asarray(A ,dtype=np.floataa ) for feature in input_features] UpperCAmelCase__ : Dict = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: UpperCAmelCase__ : Union[str, Any] = ( np.array(A ,dtype=np.intaa ) if self._get_padding_strategies(A ,max_length=A ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) UpperCAmelCase__ : Any = self.normalize( padded_inputs["""input_features"""] ,attention_mask=A ) if return_tensors is not None: UpperCAmelCase__ : Union[str, Any] = padded_inputs.convert_to_tensors(A ) return padded_inputs
65
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = {'configuration_timm_backbone': ['TimmBackboneConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['TimmBackbone'] if TYPE_CHECKING: from .configuration_timm_backbone import TimmBackboneConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timm_backbone import TimmBackbone else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" from math import factorial def lowerCAmelCase ( __UpperCamelCase = 100 ): '''simple docstring''' return sum(int(__UpperCamelCase ) for x in str(factorial(__UpperCamelCase ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
65
1
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): def __init__( self : Union[str, Any] ,A : List[Any]=None ,**A : Any ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" ,A ,) super().__init__(args=A ,**A )
65
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Optional[int] ,A : int=13 ,A : Tuple=7 ,A : Dict=True ,A : Optional[int]=True ,A : Tuple=True ,A : str=True ,A : Any=99 ,A : Tuple=32 ,A : Dict=5 ,A : Optional[int]=4 ,A : Dict=37 ,A : Any="gelu" ,A : Any=0.1 ,A : Optional[int]=0.1 ,A : Union[str, Any]=512 ,A : Any=16 ,A : List[str]=2 ,A : List[Any]=0.0_2 ,A : Optional[int]=4 ,): '''simple docstring''' UpperCAmelCase__ : Dict = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : Optional[int] = is_training UpperCAmelCase__ : Optional[Any] = use_attention_mask UpperCAmelCase__ : int = use_token_type_ids UpperCAmelCase__ : int = use_labels UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : List[Any] = type_vocab_size UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : List[Any] = num_choices def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : List[str] = None if self.use_attention_mask: UpperCAmelCase__ : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : int = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=A ,) return config, input_ids, attention_mask def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = config_and_inputs UpperCAmelCase__ : str = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = FlaxDistilBertModelTester(self ) @slow def __lowercase ( self : Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = model_class_name.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(A ) @require_flax class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) UpperCAmelCase__ : str = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ : Dict = model(A ,attention_mask=A )[0] UpperCAmelCase__ : List[Any] = (1, 11, 768) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Any = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,A ,atol=1e-4 ) )
65
1
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __UpperCAmelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.14.0', 'To fix: pip install -r examples/pytorch/audio-classification/requirements.txt') def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 16000 ): '''simple docstring''' UpperCAmelCase__ : int = int(round(sample_rate * max_length ) ) if len(__UpperCamelCase ) <= sample_length: return wav UpperCAmelCase__ : List[str] = randint(0 , len(__UpperCamelCase ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class __lowercase : snake_case_ = field(default=__lowerCamelCase , metadata={"""help""": """Name of a dataset from the datasets package"""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """A file containing the training audio paths and labels."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """A file containing the validation audio paths and labels."""} ) snake_case_ = field( default="""train""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) snake_case_ = field( default="""validation""" , metadata={ """help""": ( """The name of the training data set split to use (via the datasets library). Defaults to 'validation'""" ) } , ) snake_case_ = field( default="""audio""" , metadata={"""help""": """The name of the dataset column containing the audio data. Defaults to 'audio'"""} , ) snake_case_ = field( default="""label""" , metadata={"""help""": """The name of the dataset column containing the labels. Defaults to 'label'"""} ) snake_case_ = field( default=__lowerCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) snake_case_ = field( default=__lowerCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) snake_case_ = field( default=2_0 , metadata={"""help""": """Audio clips will be randomly cut to this length during training if the value is set."""} , ) @dataclass class __lowercase : snake_case_ = field( default="""facebook/wav2vec2-base""" , metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} , ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from the Hub"""} ) snake_case_ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Name or path of preprocessor config."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Whether to freeze the feature encoder layers of the model."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Whether to generate an attention mask in the feature extractor."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Will enable to load a pretrained model whose head dimensions are different."""} , ) def __lowercase ( self : Dict ): '''simple docstring''' if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( """The argument `--freeze_feature_extractor` is deprecated and """ """will be removed in a future version. Use `--freeze_feature_encoder`""" """instead. Setting `freeze_feature_encoder==True`.""" ,A ,) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( """The argument `--freeze_feature_extractor` is deprecated and """ """should not be used in combination with `--freeze_feature_encoder`.""" """Only make use of `--freeze_feature_encoder`.""" ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_audio_classification""" , __UpperCamelCase , __UpperCamelCase ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase__ : Any = training_args.get_process_log_level() logger.setLevel(__UpperCamelCase ) transformers.utils.logging.set_verbosity(__UpperCamelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} " + F"distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}" ) logger.info(F"Training/evaluation parameters {training_args}" ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. UpperCAmelCase__ : Optional[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase__ : Optional[Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. " """Use --overwrite_output_dir to train from scratch.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change " """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset and prepare it for the audio classification task. UpperCAmelCase__ : Union[str, Any] = DatasetDict() UpperCAmelCase__ : Tuple = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) UpperCAmelCase__ : Optional[int] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( F"--audio_column_name {data_args.audio_column_name} not found in dataset '{data_args.dataset_name}'. " """Make sure to set `--audio_column_name` to the correct audio column - one of """ F"{', '.join(raw_datasets['train'].column_names )}." ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( F"--label_column_name {data_args.label_column_name} not found in dataset '{data_args.dataset_name}'. " """Make sure to set `--label_column_name` to the correct text column - one of """ F"{', '.join(raw_datasets['train'].column_names )}." ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy UpperCAmelCase__ : List[Any] = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. UpperCAmelCase__ : List[Any] = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) UpperCAmelCase__ : Dict = feature_extractor.model_input_names[0] def train_transforms(__UpperCamelCase ): UpperCAmelCase__ : List[str] = [] for audio in batch[data_args.audio_column_name]: UpperCAmelCase__ : Dict = random_subsample( audio["""array"""] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(__UpperCamelCase ) UpperCAmelCase__ : List[str] = feature_extractor(__UpperCamelCase , sampling_rate=feature_extractor.sampling_rate ) UpperCAmelCase__ : List[str] = {model_input_name: inputs.get(__UpperCamelCase )} UpperCAmelCase__ : Tuple = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(__UpperCamelCase ): UpperCAmelCase__ : str = [audio["""array"""] for audio in batch[data_args.audio_column_name]] UpperCAmelCase__ : List[Any] = feature_extractor(__UpperCamelCase , sampling_rate=feature_extractor.sampling_rate ) UpperCAmelCase__ : Dict = {model_input_name: inputs.get(__UpperCamelCase )} UpperCAmelCase__ : List[str] = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. UpperCAmelCase__ : Any = raw_datasets["""train"""].features[data_args.label_column_name].names UpperCAmelCase__ , UpperCAmelCase__ : List[str] = {}, {} for i, label in enumerate(__UpperCamelCase ): UpperCAmelCase__ : Any = str(__UpperCamelCase ) UpperCAmelCase__ : int = label # Load the accuracy metric from the datasets package UpperCAmelCase__ : Any = evaluate.load("""accuracy""" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(__UpperCamelCase ): UpperCAmelCase__ : Dict = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=__UpperCamelCase , references=eval_pred.label_ids ) UpperCAmelCase__ : Any = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__UpperCamelCase ) , labelaid=__UpperCamelCase , idalabel=__UpperCamelCase , finetuning_task="""audio-classification""" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) UpperCAmelCase__ : List[Any] = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=__UpperCamelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: UpperCAmelCase__ : Optional[int] = ( raw_datasets["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(__UpperCamelCase , output_all_columns=__UpperCamelCase ) if training_args.do_eval: if data_args.max_eval_samples is not None: UpperCAmelCase__ : Tuple = ( raw_datasets["""eval"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(__UpperCamelCase , output_all_columns=__UpperCamelCase ) # Initialize our trainer UpperCAmelCase__ : Optional[Any] = Trainer( model=__UpperCamelCase , args=__UpperCamelCase , train_dataset=raw_datasets["""train"""] if training_args.do_train else None , eval_dataset=raw_datasets["""eval"""] if training_args.do_eval else None , compute_metrics=__UpperCamelCase , tokenizer=__UpperCamelCase , ) # Training if training_args.do_train: UpperCAmelCase__ : Dict = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase__ : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase__ : str = last_checkpoint UpperCAmelCase__ : str = trainer.train(resume_from_checkpoint=__UpperCamelCase ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase__ : int = trainer.evaluate() trainer.log_metrics("""eval""" , __UpperCamelCase ) trainer.save_metrics("""eval""" , __UpperCamelCase ) # Write model card and (optionally) push to hub UpperCAmelCase__ : List[str] = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """audio-classification""", """dataset""": data_args.dataset_name, """tags""": ["""audio-classification"""], } if training_args.push_to_hub: trainer.push_to_hub(**__UpperCamelCase ) else: trainer.create_model_card(**__UpperCamelCase ) if __name__ == "__main__": main()
65
"""simple docstring""" __UpperCAmelCase = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image', 'mask_image']) __UpperCAmelCase = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['example_image', 'image', 'mask_image']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset(['input_tokens']) __UpperCAmelCase = frozenset(['input_tokens'])
65
1
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase=False ): '''simple docstring''' UpperCAmelCase__ : Any = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"module.blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"module.blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append( (F"module.blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"module.blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"module.blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"module.blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" UpperCAmelCase__ : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase__ : int = """""" else: UpperCAmelCase__ : List[str] = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase__ : Dict = state_dict.pop(F"module.blocks.{i}.attn.qkv.weight" ) UpperCAmelCase__ : Union[str, Any] = state_dict.pop(F"module.blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase__ : Dict = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase__ : List[str] = in_proj_bias[: config.hidden_size] UpperCAmelCase__ : Dict = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase__ : str = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase__ : Optional[int] = in_proj_bias[-config.hidden_size :] def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = dct.pop(__UpperCamelCase ) UpperCAmelCase__ : Tuple = val def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = ViTMSNConfig() UpperCAmelCase__ : Any = 1000 UpperCAmelCase__ : List[Any] = """datasets/huggingface/label-files""" UpperCAmelCase__ : int = """imagenet-1k-id2label.json""" UpperCAmelCase__ : List[Any] = json.load(open(hf_hub_download(__UpperCamelCase , __UpperCamelCase ) , """r""" ) ) UpperCAmelCase__ : Tuple = {int(__UpperCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase__ : Optional[int] = idalabel UpperCAmelCase__ : Any = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: UpperCAmelCase__ : Dict = 384 UpperCAmelCase__ : int = 1536 UpperCAmelCase__ : Union[str, Any] = 6 elif "l16" in checkpoint_url: UpperCAmelCase__ : str = 1024 UpperCAmelCase__ : List[Any] = 4096 UpperCAmelCase__ : Union[str, Any] = 24 UpperCAmelCase__ : Tuple = 16 UpperCAmelCase__ : Dict = 0.1 elif "b4" in checkpoint_url: UpperCAmelCase__ : Union[str, Any] = 4 elif "l7" in checkpoint_url: UpperCAmelCase__ : Optional[Any] = 7 UpperCAmelCase__ : Tuple = 1024 UpperCAmelCase__ : Dict = 4096 UpperCAmelCase__ : List[Any] = 24 UpperCAmelCase__ : Union[str, Any] = 16 UpperCAmelCase__ : Dict = 0.1 UpperCAmelCase__ : int = ViTMSNModel(__UpperCamelCase ) UpperCAmelCase__ : Any = torch.hub.load_state_dict_from_url(__UpperCamelCase , map_location="""cpu""" )["""target_encoder"""] UpperCAmelCase__ : List[str] = ViTImageProcessor(size=config.image_size ) remove_projection_head(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = create_rename_keys(__UpperCamelCase , base_model=__UpperCamelCase ) for src, dest in rename_keys: rename_key(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) read_in_q_k_v(__UpperCamelCase , __UpperCamelCase , base_model=__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) model.eval() UpperCAmelCase__ : Any = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase__ : List[Any] = Image.open(requests.get(__UpperCamelCase , stream=__UpperCamelCase ).raw ) UpperCAmelCase__ : Tuple = ViTImageProcessor( size=config.image_size , image_mean=__UpperCamelCase , image_std=__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = image_processor(images=__UpperCamelCase , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) UpperCAmelCase__ : Any = model(**__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: UpperCAmelCase__ : int = torch.tensor([[-1.0915, -1.4876, -1.1809]] ) elif "b16" in checkpoint_url: UpperCAmelCase__ : Optional[Any] = torch.tensor([[14.2889, -18.9045, 11.7281]] ) elif "l16" in checkpoint_url: UpperCAmelCase__ : int = torch.tensor([[41.5028, -22.8681, 45.6475]] ) elif "b4" in checkpoint_url: UpperCAmelCase__ : Dict = torch.tensor([[-4.3868, 5.2932, -0.4137]] ) else: UpperCAmelCase__ : Optional[Any] = torch.tensor([[-0.1792, -0.6465, 2.4263]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , __UpperCamelCase , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__UpperCamelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) __UpperCAmelCase = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
65
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase__ : Tuple = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[Any] = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase__ : int = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 16_000, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase__ : Optional[int] = tempfile.mkdtemp() UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,A ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.feature_extraction_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) # load decoder from hub UpperCAmelCase__ : int = """hf-internal-testing/ngram-beam-search-decoder""" def __lowercase ( self : str ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.add_kwargs_tokens_map.copy() kwargs.update(A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[str] ,**A : Dict ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Any ,**A : List[Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name ,**A ) def __lowercase ( self : Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_decoder() UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() ,feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor ,A ) # decoder self.assertEqual(processor.decoder._alphabet.labels ,decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set ,decoder.model_container[decoder._model_key]._unigram_set ,) self.assertIsInstance(processor.decoder ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname ,alpha=5.0 ,beta=3.0 ,score_boundary=-7.0 ,unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha ,5.0 ) self.assertEqual(processor.language_model.beta ,3.0 ) self.assertEqual(processor.language_model.score_boundary ,-7.0 ) self.assertEqual(processor.language_model.unk_score_offset ,3 ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(A ,"""include""" ): WavaVecaProcessorWithLM( tokenizer=A ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : Optional[Any] = feature_extractor(A ,return_tensors="""np""" ) UpperCAmelCase__ : List[Any] = processor(A ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : List[Any] = """This is a test string""" UpperCAmelCase__ : int = processor(text=A ) UpperCAmelCase__ : Dict = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __lowercase ( self : Tuple ,A : List[Any]=(2, 10, 16) ,A : Dict=77 ): '''simple docstring''' np.random.seed(A ) return np.random.rand(*A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits(shape=(10, 16) ,seed=13 ) UpperCAmelCase__ : Tuple = processor.decode(A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(A )[0] self.assertEqual(decoded_decoder[0] ,decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" ,decoded_processor.text ) self.assertEqual(decoded_decoder[-2] ,decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] ,decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def __lowercase ( self : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_feature_extractor() UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : List[str] = processor.batch_decode(A ) else: with get_context(A ).Pool() as pool: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(A ,A ) UpperCAmelCase__ : Optional[Any] = list(A ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A ,decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] ,decoded_processor.text ) self.assertListEqual(A ,decoded_processor.logit_score ) self.assertListEqual(A ,decoded_processor.lm_score ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits() UpperCAmelCase__ : Any = 15 UpperCAmelCase__ : Dict = -2_0.0 UpperCAmelCase__ : List[Any] = -4.0 UpperCAmelCase__ : Union[str, Any] = processor.batch_decode( A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[str] = decoded_processor_out.text UpperCAmelCase__ : List[str] = list(A ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( A ,A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] ,A ) self.assertTrue(np.array_equal(A ,decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] ,A ,atol=1e-3 ) ) self.assertTrue(np.array_equal(A ,decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] ,A ,atol=1e-3 ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits() UpperCAmelCase__ : Tuple = 2.0 UpperCAmelCase__ : str = 5.0 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = processor.batch_decode( A ,alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) UpperCAmelCase__ : Any = decoded_processor_out.text UpperCAmelCase__ : Union[str, Any] = list(A ) decoder.reset_params( alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch( A ,A ,) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] ,A ) UpperCAmelCase__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha ,2.0 ) self.assertEqual(lm_model.beta ,5.0 ) self.assertEqual(lm_model.unk_score_offset ,-2_0.0 ) self.assertEqual(lm_model.score_boundary ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : str = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Any = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[int] = os.listdir(A ) UpperCAmelCase__ : List[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(A ) UpperCAmelCase__ : Tuple = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[int] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Tuple = os.listdir(A ) UpperCAmelCase__ : Dict = os.listdir(A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = floats_list((3, 1_000) ) UpperCAmelCase__ : List[str] = processor_wavaveca(A ,return_tensors="""np""" ) UpperCAmelCase__ : Dict = processor_auto(A ,return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() ,input_auto[key].sum() ,delta=1e-2 ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Tuple = processor_wavaveca.batch_decode(A ) UpperCAmelCase__ : List[str] = processor_auto.batch_decode(A ) self.assertListEqual(decoded_wavaveca.text ,decoded_auto.text ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : int = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) self.assertListEqual( processor.model_input_names ,feature_extractor.model_input_names ,msg="""`processor` and `feature_extractor` model input names do not match""" ,) @staticmethod def __lowercase ( A : Optional[Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = self._get_dummy_logits()[0] UpperCAmelCase__ : List[str] = processor.decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ) ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""end_offset""" ) ,[1, 3, 5] ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : int = self._get_dummy_logits() UpperCAmelCase__ : Any = processor.batch_decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(A ,"""word""" ) ) for o in outputs["""word_offsets"""]] ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""end_offset""" ) ,[1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self : Tuple ): '''simple docstring''' import torch UpperCAmelCase__ : Any = load_dataset("""common_voice""" ,"""en""" ,split="""train""" ,streaming=A ) UpperCAmelCase__ : Tuple = ds.cast_column("""audio""" ,datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Tuple = iter(A ) UpperCAmelCase__ : Optional[int] = next(A ) UpperCAmelCase__ : List[Any] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase__ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Tuple = processor(sample["""audio"""]["""array"""] ,return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(A ).logits.cpu().numpy() UpperCAmelCase__ : Any = processor.decode(logits[0] ,output_word_offsets=A ) UpperCAmelCase__ : str = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase__ : Dict = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,A ) self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,output.text ) # output times UpperCAmelCase__ : str = torch.tensor(self.get_from_offsets(A ,"""start_time""" ) ) UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(A ,"""end_time""" ) ) # fmt: off UpperCAmelCase__ : Union[str, Any] = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : List[Any] = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) ) self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) )
65
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class __lowercase : def __init__( self : Any ,A : Union[str, Any] ,A : Optional[Any]=2 ,A : Tuple=True ,A : Tuple=False ,A : Optional[int]=10 ,A : Any=3 ,A : Tuple=32 * 8 ,A : List[Any]=32 * 8 ,A : int=4 ,A : List[Any]=64 ,): '''simple docstring''' UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : Tuple = is_training UpperCAmelCase__ : Optional[Any] = use_auxiliary_loss UpperCAmelCase__ : int = num_queries UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : List[str] = min_size UpperCAmelCase__ : Optional[Any] = max_size UpperCAmelCase__ : Tuple = num_labels UpperCAmelCase__ : List[str] = hidden_dim UpperCAmelCase__ : Union[str, Any] = hidden_dim def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( A ) UpperCAmelCase__ : Optional[int] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=A ) UpperCAmelCase__ : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=A ) > 0.5 ).float() UpperCAmelCase__ : Optional[Any] = (torch.rand((self.batch_size, self.num_labels) ,device=A ) > 0.5).long() UpperCAmelCase__ : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MaskaFormerConfig( hidden_size=self.hidden_dim ,) UpperCAmelCase__ : int = self.num_queries UpperCAmelCase__ : str = self.num_labels UpperCAmelCase__ : List[Any] = [1, 1, 1, 1] UpperCAmelCase__ : List[Any] = self.num_channels UpperCAmelCase__ : List[Any] = 64 UpperCAmelCase__ : str = 128 UpperCAmelCase__ : int = self.hidden_dim UpperCAmelCase__ : List[Any] = self.hidden_dim UpperCAmelCase__ : int = self.hidden_dim return config def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = self.prepare_config_and_inputs() UpperCAmelCase__ : List[Any] = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def __lowercase ( self : Optional[int] ,A : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = output.encoder_hidden_states UpperCAmelCase__ : List[str] = output.pixel_decoder_hidden_states UpperCAmelCase__ : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(A ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(A ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(A ) ,config.decoder_layers ) def __lowercase ( self : List[Any] ,A : List[Any] ,A : Dict ,A : Union[str, Any] ,A : str=False ): '''simple docstring''' with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = MaskaFormerModel(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : Tuple = model(pixel_values=A ,pixel_mask=A ) UpperCAmelCase__ : Optional[int] = model(A ,output_hidden_states=A ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(A ,A ) def __lowercase ( self : Tuple ,A : List[str] ,A : Dict ,A : Tuple ,A : Any ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaskaFormerForUniversalSegmentation(config=A ) model.to(A ) model.eval() def comm_check_on_output(A : str ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = model(pixel_values=A ,pixel_mask=A ) UpperCAmelCase__ : Tuple = model(A ) comm_check_on_output(A ) UpperCAmelCase__ : Optional[Any] = model( pixel_values=A ,pixel_mask=A ,mask_labels=A ,class_labels=A ) comm_check_on_output(A ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class __lowercase ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): snake_case_ = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () snake_case_ = {"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaskaFormerModelTester(self ) UpperCAmelCase__ : Union[str, Any] = ConfigTester(self ,config_class=A ,has_text_modality=A ) def __lowercase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(A ,**A ,output_hidden_states=A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*A ) @unittest.skip(reason="""Mask2Former does not use inputs_embeds""" ) def __lowercase ( self : List[str] ): '''simple docstring''' pass @unittest.skip(reason="""Mask2Former does not have a get_input_embeddings method""" ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Mask2Former is not a generative model""" ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Mask2Former does not use token embeddings""" ) def __lowercase ( self : str ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason="""Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def __lowercase ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : int = model_class(A ) UpperCAmelCase__ : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : List[str] = [*signature.parameters.keys()] UpperCAmelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,A ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: UpperCAmelCase__ : Union[str, Any] = MaskaFormerModel.from_pretrained(A ) self.assertIsNotNone(A ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = (self.model_tester.min_size,) * 2 UpperCAmelCase__ : Dict = { """pixel_values""": torch.randn((2, 3, *size) ,device=A ), """mask_labels""": torch.randn((2, 10, *size) ,device=A ), """class_labels""": torch.zeros(2 ,10 ,device=A ).long(), } UpperCAmelCase__ : List[Any] = self.model_tester.get_config() UpperCAmelCase__ : Union[str, Any] = MaskaFormerForUniversalSegmentation(A ).to(A ) UpperCAmelCase__ : List[str] = model(**A ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(A ,**A ,output_hidden_states=A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(A ).to(A ) UpperCAmelCase__ : str = model(**A ,output_attentions=A ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Optional[int] ): '''simple docstring''' if not self.model_tester.is_training: return UpperCAmelCase__ : Optional[int] = self.all_model_classes[1] UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() UpperCAmelCase__ : Union[str, Any] = model_class(A ) model.to(A ) model.train() UpperCAmelCase__ : Optional[Any] = model(A ,mask_labels=A ,class_labels=A ).loss loss.backward() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.all_model_classes[1] UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : int = model_class(A ).to(A ) model.train() UpperCAmelCase__ : Tuple = model(A ,mask_labels=A ,class_labels=A ) UpperCAmelCase__ : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCAmelCase__ : Tuple = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() UpperCAmelCase__ : int = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCAmelCase__ : Union[str, Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=A ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __UpperCAmelCase = 1E-4 def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class __lowercase ( unittest.TestCase ): @cached_property def __lowercase ( self : str ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : str ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(A ) UpperCAmelCase__ : str = self.default_image_processor UpperCAmelCase__ : List[Any] = prepare_img() UpperCAmelCase__ : Tuple = image_processor(A ,return_tensors="""pt""" ).to(A ) UpperCAmelCase__ : Optional[Any] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(A ,(1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase__ : int = model(**A ) UpperCAmelCase__ : List[Any] = torch.tensor( [[-0.2_7_9_0, -1.0_7_1_7, -1.1_6_6_8], [-0.5_1_2_8, -0.3_1_2_8, -0.4_9_8_7], [-0.5_8_3_2, 0.1_9_7_1, -0.0_1_9_7]] ).to(A ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,A ,atol=A ) ) UpperCAmelCase__ : Union[str, Any] = torch.tensor( [[0.8_9_7_3, 1.1_8_4_7, 1.1_7_7_6], [1.1_9_3_4, 1.5_0_4_0, 1.5_1_2_8], [1.1_1_5_3, 1.4_4_8_6, 1.4_9_5_1]] ).to(A ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,A ,atol=A ) ) UpperCAmelCase__ : int = torch.tensor( [[2.1_1_5_2, 1.7_0_0_0, -0.8_6_0_3], [1.5_8_0_8, 1.8_0_0_4, -0.9_3_5_3], [1.6_0_4_3, 1.7_4_9_5, -0.5_9_9_9]] ).to(A ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,A ,atol=A ) ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : int = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(A ).eval() UpperCAmelCase__ : Union[str, Any] = self.default_image_processor UpperCAmelCase__ : List[str] = prepare_img() UpperCAmelCase__ : List[str] = image_processor(A ,return_tensors="""pt""" ).to(A ) UpperCAmelCase__ : Optional[int] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(A ,(1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = model(**A ) # masks_queries_logits UpperCAmelCase__ : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) UpperCAmelCase__ : Optional[Any] = [ [-8.7_8_3_9, -9.0_0_5_6, -8.8_1_2_1], [-7.4_1_0_4, -7.0_3_1_3, -6.5_4_0_1], [-6.6_1_0_5, -6.3_4_2_7, -6.4_6_7_5], ] UpperCAmelCase__ : Dict = torch.tensor(A ).to(A ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,A ,atol=A ) ) # class_queries_logits UpperCAmelCase__ : Any = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase__ : str = torch.tensor( [ [1.8_3_2_4, -8.0_8_3_5, -4.1_9_2_2], [0.8_4_5_0, -9.0_0_5_0, -3.6_0_5_3], [0.3_0_4_5, -7.7_2_9_3, -3.0_2_7_5], ] ).to(A ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,A ,atol=A ) ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(A ).eval() UpperCAmelCase__ : str = self.default_image_processor UpperCAmelCase__ : Any = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors="""pt""" ,) UpperCAmelCase__ : Tuple = inputs["""pixel_values"""].to(A ) UpperCAmelCase__ : List[Any] = [el.to(A ) for el in inputs["""mask_labels"""]] UpperCAmelCase__ : Union[str, Any] = [el.to(A ) for el in inputs["""class_labels"""]] with torch.no_grad(): UpperCAmelCase__ : Optional[int] = model(**A ) self.assertTrue(outputs.loss is not None )
65
"""simple docstring""" from sklearn.metrics import fa_score import datasets __UpperCAmelCase = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __UpperCAmelCase = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __UpperCAmelCase = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def __lowercase ( self : List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) ,reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"""] ,) def __lowercase ( self : Union[str, Any] ,A : List[str] ,A : List[Any] ,A : Optional[Any]=None ,A : List[str]=1 ,A : Optional[Any]="binary" ,A : Any=None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = fa_score( A ,A ,labels=A ,pos_label=A ,average=A ,sample_weight=A ) return {"f1": float(A ) if score.size == 1 else score}
65
1
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """attention_mask"""] def __init__( self : Any ,A : str=80 ,A : Optional[int]=16_000 ,A : int=0.0 ,A : str=10 ,A : Any=25 ,A : str="hamming_window" ,A : int=3_2_7_6_8.0 ,A : List[str]=0.9_7 ,A : Optional[int]=1.0 ,A : Optional[Any]=True ,A : Tuple=True ,A : Any=False ,**A : int ,): '''simple docstring''' super().__init__(feature_size=A ,sampling_rate=A ,padding_value=A ,**A ) UpperCAmelCase__ : str = feature_size UpperCAmelCase__ : int = sampling_rate UpperCAmelCase__ : int = padding_value UpperCAmelCase__ : Dict = hop_length UpperCAmelCase__ : int = win_length UpperCAmelCase__ : Dict = frame_signal_scale UpperCAmelCase__ : Dict = preemphasis_coeff UpperCAmelCase__ : str = mel_floor UpperCAmelCase__ : Any = normalize_means UpperCAmelCase__ : str = normalize_vars UpperCAmelCase__ : int = win_function UpperCAmelCase__ : List[Any] = return_attention_mask UpperCAmelCase__ : str = win_length * sampling_rate // 1_000 UpperCAmelCase__ : List[Any] = hop_length * sampling_rate // 1_000 UpperCAmelCase__ : int = optimal_fft_length(self.sample_size ) UpperCAmelCase__ : List[Any] = (self.n_fft // 2) + 1 def __lowercase ( self : Union[str, Any] ,A : np.array ): '''simple docstring''' if self.win_function == "hamming_window": UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=A ) else: UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ) UpperCAmelCase__ : Union[str, Any] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) UpperCAmelCase__ : Optional[Any] = spectrogram( one_waveform * self.frame_signal_scale ,window=A ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=A ,preemphasis=self.preemphasis_coeff ,mel_filters=A ,mel_floor=self.mel_floor ,log_mel="""log""" ,) return msfc_features.T def __lowercase ( self : str ,A : Any ,A : Optional[int] ,A : str ): '''simple docstring''' # make sure we normalize float32 arrays if self.normalize_means: UpperCAmelCase__ : Optional[Any] = x[:input_length].mean(axis=0 ) UpperCAmelCase__ : Any = np.subtract(A ,A ) if self.normalize_vars: UpperCAmelCase__ : str = x[:input_length].std(axis=0 ) UpperCAmelCase__ : Optional[int] = np.divide(A ,A ) if input_length < x.shape[0]: UpperCAmelCase__ : int = padding_value # make sure array is in float32 UpperCAmelCase__ : str = x.astype(np.floataa ) return x def __lowercase ( self : Union[str, Any] ,A : List[np.ndarray] ,A : Optional[np.ndarray] = None ): '''simple docstring''' UpperCAmelCase__ : Any = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(A ,A ,self.padding_value ) for x, n in zip(A ,A )] def __call__( self : Union[str, Any] ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : Union[bool, str, PaddingStrategy] = False ,A : Optional[int] = None ,A : bool = False ,A : Optional[int] = None ,A : Optional[bool] = None ,A : Optional[Union[str, TensorType]] = None ,A : Optional[int] = None ,**A : Tuple ,): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" f" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" f" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[Any] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : Any = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : List[str] = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Union[str, Any] = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [raw_speech] # extract fbank features UpperCAmelCase__ : Tuple = [self._extract_mfsc_features(A ) for one_waveform in raw_speech] # convert into correct format for padding UpperCAmelCase__ : str = BatchFeature({"""input_features""": features} ) UpperCAmelCase__ : Optional[Any] = self.pad( A ,padding=A ,max_length=A ,truncation=A ,pad_to_multiple_of=A ,return_attention_mask=A ,**A ,) # make sure list is in array format UpperCAmelCase__ : Tuple = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] ,A ): UpperCAmelCase__ : Union[str, Any] = [np.asarray(A ,dtype=np.floataa ) for feature in input_features] UpperCAmelCase__ : Dict = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: UpperCAmelCase__ : Union[str, Any] = ( np.array(A ,dtype=np.intaa ) if self._get_padding_strategies(A ,max_length=A ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) UpperCAmelCase__ : Any = self.normalize( padded_inputs["""input_features"""] ,attention_mask=A ) if return_tensors is not None: UpperCAmelCase__ : Union[str, Any] = padded_inputs.convert_to_tensors(A ) return padded_inputs
65
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece.model') __UpperCAmelCase = {'target_lang': 'fi', 'source_lang': 'en'} __UpperCAmelCase = '>>zh<<' __UpperCAmelCase = 'Helsinki-NLP/' if is_torch_available(): __UpperCAmelCase = 'pt' elif is_tf_available(): __UpperCAmelCase = 'tf' else: __UpperCAmelCase = 'jax' @require_sentencepiece class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = MarianTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Optional[int] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Optional[Any] = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] UpperCAmelCase__ : int = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[int] = Path(self.tmpdirname ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(A ,save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(A ,save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) UpperCAmelCase__ : Dict = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self : List[Any] ,**A : List[Any] ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Union[str, Any] ,A : Tuple ): '''simple docstring''' return ( "This is a test", "This is a test", ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """</s>""" UpperCAmelCase__ : int = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""</s>""" ) self.assertEqual(vocab_keys[1] ,"""<unk>""" ) self.assertEqual(vocab_keys[-1] ,"""<pad>""" ) self.assertEqual(len(A ) ,9 ) def __lowercase ( self : Dict ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,9 ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MarianTokenizer.from_pretrained(f"{ORG_NAME}opus-mt-en-de" ) UpperCAmelCase__ : List[str] = en_de_tokenizer(["""I am a small frog"""] ,return_tensors=A ) self.assertIsInstance(A ,A ) UpperCAmelCase__ : str = [38, 121, 14, 697, 38_848, 0] self.assertListEqual(A ,batch.input_ids[0] ) UpperCAmelCase__ : Optional[Any] = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(A ) UpperCAmelCase__ : Tuple = [x.name for x in Path(A ).glob("""*""" )] self.assertIn("""source.spm""" ,A ) MarianTokenizer.from_pretrained(A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = tok( ["""I am a small frog""" * 1_000, """I am a small frog"""] ,padding=A ,truncation=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch.input_ids.shape ,(2, 512) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : Tuple = tok(["""I am a tiny frog""", """I am a small frog"""] ,padding=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch_smaller.input_ids.shape ,(2, 10) ) @slow def __lowercase ( self : Dict ): '''simple docstring''' # fmt: off UpperCAmelCase__ : Optional[int] = {"""input_ids""": [[43_495, 462, 20, 42_164, 1_369, 52, 464, 132, 1_703, 492, 13, 7_491, 38_999, 6, 8, 464, 132, 1_703, 492, 13, 4_669, 37_867, 13, 7_525, 27, 1_593, 988, 13, 33_972, 7_029, 6, 20, 8_251, 383, 2, 270, 5_866, 3_788, 2, 2_353, 8_251, 12_338, 2, 13_958, 387, 2, 3_629, 6_953, 188, 2_900, 2, 13_958, 8_011, 11_501, 23, 8_460, 4_073, 34_009, 20, 435, 11_439, 27, 8, 8_460, 4_073, 6_004, 20, 9_988, 375, 27, 33, 266, 1_945, 1_076, 1_350, 37_867, 3_288, 5, 577, 1_076, 4_374, 8, 5_082, 5, 26_453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10_767, 6, 316, 304, 4_239, 3, 0], [148, 15_722, 19, 1_839, 12, 1_350, 13, 22_327, 5_082, 5_418, 47_567, 35_938, 59, 318, 19_552, 108, 2_183, 54, 14_976, 4_835, 32, 547, 1_114, 8, 315, 2_417, 5, 92, 19_088, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100], [36, 6_395, 12_570, 39_147, 11_597, 6, 266, 4, 45_405, 7_296, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""Helsinki-NLP/opus-mt-en-de""" ,revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" ,decode_kwargs={"""use_source_tokenizer""": True} ,) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) UpperCAmelCase__ : Any = """Tämä on testi""" UpperCAmelCase__ : int = """This is a test""" UpperCAmelCase__ : List[str] = [76, 7, 2_047, 2] UpperCAmelCase__ : Optional[Any] = [69, 12, 11, 940, 2] UpperCAmelCase__ : List[str] = tokenizer(A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : Optional[int] = tokenizer(text_target=A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : int = tokenizer.decode(A ,skip_special_tokens=A ) self.assertEqual(A ,A )
65
1
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Optional[int] ,A : int=13 ,A : Tuple=7 ,A : Dict=True ,A : Optional[int]=True ,A : Tuple=True ,A : str=True ,A : Any=99 ,A : Tuple=32 ,A : Dict=5 ,A : Optional[int]=4 ,A : Dict=37 ,A : Any="gelu" ,A : Any=0.1 ,A : Optional[int]=0.1 ,A : Union[str, Any]=512 ,A : Any=16 ,A : List[str]=2 ,A : List[Any]=0.0_2 ,A : Optional[int]=4 ,): '''simple docstring''' UpperCAmelCase__ : Dict = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : Optional[int] = is_training UpperCAmelCase__ : Optional[Any] = use_attention_mask UpperCAmelCase__ : int = use_token_type_ids UpperCAmelCase__ : int = use_labels UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : List[Any] = type_vocab_size UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : List[Any] = num_choices def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : List[str] = None if self.use_attention_mask: UpperCAmelCase__ : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : int = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=A ,) return config, input_ids, attention_mask def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = config_and_inputs UpperCAmelCase__ : str = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = FlaxDistilBertModelTester(self ) @slow def __lowercase ( self : Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = model_class_name.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(A ) @require_flax class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) UpperCAmelCase__ : str = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ : Dict = model(A ,attention_mask=A )[0] UpperCAmelCase__ : List[Any] = (1, 11, 768) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Any = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,A ,atol=1e-4 ) )
65
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=__lowerCamelCase ): snake_case_ = ["""onnx"""] def __init__( self : int ,*A : List[str] ,**A : int ): '''simple docstring''' requires_backends(self ,["""onnx"""] ) @classmethod def __lowercase ( cls : Optional[Any] ,*A : List[str] ,**A : Dict ): '''simple docstring''' requires_backends(cls ,["""onnx"""] ) @classmethod def __lowercase ( cls : List[Any] ,*A : Optional[int] ,**A : int ): '''simple docstring''' requires_backends(cls ,["""onnx"""] )
65
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} __UpperCAmelCase = { 'vocab_file': { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json' ), 'distilbert-base-german-cased': ( 'https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json' ), 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json' ), }, } __UpperCAmelCase = { 'distilbert-base-uncased': 512, 'distilbert-base-uncased-distilled-squad': 512, 'distilbert-base-cased': 512, 'distilbert-base-cased-distilled-squad': 512, 'distilbert-base-german-cased': 512, 'distilbert-base-multilingual-cased': 512, } __UpperCAmelCase = { 'distilbert-base-uncased': {'do_lower_case': True}, 'distilbert-base-uncased-distilled-squad': {'do_lower_case': True}, 'distilbert-base-cased': {'do_lower_case': False}, 'distilbert-base-cased-distilled-squad': {'do_lower_case': False}, 'distilbert-base-german-cased': {'do_lower_case': False}, 'distilbert-base-multilingual-cased': {'do_lower_case': False}, } class __lowercase ( __lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = PRETRAINED_INIT_CONFIGURATION snake_case_ = ["""input_ids""", """attention_mask"""] snake_case_ = DistilBertTokenizer def __init__( self : List[str] ,A : str=None ,A : int=None ,A : Optional[int]=True ,A : Optional[Any]="[UNK]" ,A : Dict="[SEP]" ,A : Any="[PAD]" ,A : List[Any]="[CLS]" ,A : List[Any]="[MASK]" ,A : List[str]=True ,A : int=None ,**A : Union[str, Any] ,): '''simple docstring''' super().__init__( A ,tokenizer_file=A ,do_lower_case=A ,unk_token=A ,sep_token=A ,pad_token=A ,cls_token=A ,mask_token=A ,tokenize_chinese_chars=A ,strip_accents=A ,**A ,) UpperCAmelCase__ : List[str] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" ,A ) != do_lower_case or normalizer_state.get("""strip_accents""" ,A ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" ,A ) != tokenize_chinese_chars ): UpperCAmelCase__ : Optional[Any] = getattr(A ,normalizer_state.pop("""type""" ) ) UpperCAmelCase__ : Any = do_lower_case UpperCAmelCase__ : Any = strip_accents UpperCAmelCase__ : Union[str, Any] = tokenize_chinese_chars UpperCAmelCase__ : Optional[Any] = normalizer_class(**A ) UpperCAmelCase__ : Dict = do_lower_case def __lowercase ( self : Optional[int] ,A : Tuple ,A : List[str]=None ): '''simple docstring''' UpperCAmelCase__ : List[str] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self : Dict ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : int = [self.sep_token_id] UpperCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self : Tuple ,A : str ,A : Optional[str] = None ): '''simple docstring''' UpperCAmelCase__ : str = self._tokenizer.model.save(A ,name=A ) return tuple(A )
65
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) UpperCAmelCase__ : List[Any] = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(__UpperCamelCase ) # Let's go UpperCAmelCase__ : int = parser.parse_args() if not hasattr(__UpperCamelCase , """func""" ): parser.print_help() exit(1 ) # Run UpperCAmelCase__ : Union[str, Any] = args.func(__UpperCamelCase ) service.run() if __name__ == "__main__": main()
65
1
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __UpperCAmelCase = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __lowercase : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : List[Any] ,A : int ,A : Optional[Any]=13 ,A : Dict=7 ,A : Dict=True ,A : Any=False ,A : Dict=99 ,A : int=32 ,A : Optional[int]=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : str=0.1 ,A : int=0.1 ,A : Optional[int]=20 ,A : Tuple=2 ,A : str=1 ,A : Optional[Any]=0 ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : int = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Union[str, Any] = pad_token_id UpperCAmelCase__ : List[str] = bos_token_id def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] ,self.vocab_size ).clip(3 ,self.vocab_size ) UpperCAmelCase__ : List[str] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) ,1 ) UpperCAmelCase__ : Any = np.concatenate([input_ids, eos_tensor] ,axis=1 ) UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : str = self.config_cls( vocab_size=self.vocab_size ,d_model=self.hidden_size ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,eos_token_ids=[2] ,bos_token_id=self.bos_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.pad_token_id ,**self.config_updates ,) UpperCAmelCase__ : Optional[Any] = prepare_pegasus_inputs_dict(A ,A ,A ) return config, inputs_dict def __lowercase ( self : Any ,A : Optional[int] ,A : str ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : Dict = model_class_name(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) ,dtype="""i4""" ) UpperCAmelCase__ : str = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Optional[int] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : int = model.decode( decoder_input_ids[:, -1:] ,A ,decoder_attention_mask=A ,past_key_values=outputs_cache.past_key_values ,decoder_position_ids=A ,) UpperCAmelCase__ : Dict = model.decode(A ,A ) UpperCAmelCase__ : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def __lowercase ( self : Optional[int] ,A : str ,A : Dict ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : str = model_class_name(A ) UpperCAmelCase__ : Any = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] ,axis=-1 ,) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : Dict = model.decode( decoder_input_ids[:, -1:] ,A ,past_key_values=outputs_cache.past_key_values ,decoder_attention_mask=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Union[str, Any] = model.decode(A ,A ,decoder_attention_mask=A ) UpperCAmelCase__ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , ): '''simple docstring''' if attention_mask is None: UpperCAmelCase__ : Union[str, Any] = np.not_equal(__UpperCamelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Tuple = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) snake_case_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = FlaxPegasusModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ) def __lowercase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(A ,A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(A ,A ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : List[Any] = self._prepare_for_class(A ,A ) UpperCAmelCase__ : int = model_class(A ) @jax.jit def encode_jitted(A : Optional[int] ,A : Union[str, Any]=None ,**A : Optional[Any] ): return model.encode(input_ids=A ,attention_mask=A ) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : int = encode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : Dict = encode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : Dict = model_class(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ,inputs_dict["""attention_mask"""] ) UpperCAmelCase__ : Dict = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(A : List[Any] ,A : Any ,A : List[Any] ): return model.decode( decoder_input_ids=A ,decoder_attention_mask=A ,encoder_outputs=A ,) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : Tuple = decode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : str = decode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""google/pegasus-large""" ,from_pt=A ) UpperCAmelCase__ : Any = np.ones((1, 1) ) UpperCAmelCase__ : Optional[Any] = model(A ) self.assertIsNotNone(A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Optional[Any] = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Union[str, Any] = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] UpperCAmelCase__ : str = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] UpperCAmelCase__ : str = tokenizer(A ,return_tensors="""np""" ,truncation=A ,max_length=512 ,padding=A ) UpperCAmelCase__ : Union[str, Any] = model.generate(**A ,num_beams=2 ).sequences UpperCAmelCase__ : int = tokenizer.batch_decode(A ,skip_special_tokens=A ) assert tgt_text == decoded
65
1
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef __UpperCAmelCase = ( 'This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' warnings.warn(__UpperCamelCase , __UpperCamelCase ) requires_backends(__UpperCamelCase , """sklearn""" ) return (preds == labels).mean() def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' warnings.warn(__UpperCamelCase , __UpperCamelCase ) requires_backends(__UpperCamelCase , """sklearn""" ) UpperCAmelCase__ : Optional[Any] = simple_accuracy(__UpperCamelCase , __UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = fa_score(y_true=__UpperCamelCase , y_pred=__UpperCamelCase ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' warnings.warn(__UpperCamelCase , __UpperCamelCase ) requires_backends(__UpperCamelCase , """sklearn""" ) UpperCAmelCase__ : Optional[int] = pearsonr(__UpperCamelCase , __UpperCamelCase )[0] UpperCAmelCase__ : Union[str, Any] = spearmanr(__UpperCamelCase , __UpperCamelCase )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' warnings.warn(__UpperCamelCase , __UpperCamelCase ) requires_backends(__UpperCamelCase , """sklearn""" ) assert len(__UpperCamelCase ) == len(__UpperCamelCase ), F"Predictions and labels have mismatched lengths {len(__UpperCamelCase )} and {len(__UpperCamelCase )}" if task_name == "cola": return {"mcc": matthews_corrcoef(__UpperCamelCase , __UpperCamelCase )} elif task_name == "sst-2": return {"acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} elif task_name == "mrpc": return acc_and_fa(__UpperCamelCase , __UpperCamelCase ) elif task_name == "sts-b": return pearson_and_spearman(__UpperCamelCase , __UpperCamelCase ) elif task_name == "qqp": return acc_and_fa(__UpperCamelCase , __UpperCamelCase ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} elif task_name == "qnli": return {"acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} elif task_name == "rte": return {"acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} elif task_name == "wnli": return {"acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} elif task_name == "hans": return {"acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} else: raise KeyError(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' warnings.warn(__UpperCamelCase , __UpperCamelCase ) requires_backends(__UpperCamelCase , """sklearn""" ) if len(__UpperCamelCase ) != len(__UpperCamelCase ): raise ValueError(F"Predictions and labels have mismatched lengths {len(__UpperCamelCase )} and {len(__UpperCamelCase )}" ) if task_name == "xnli": return {"acc": simple_accuracy(__UpperCamelCase , __UpperCamelCase )} else: raise KeyError(__UpperCamelCase )
65
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) UpperCAmelCase__ : Union[str, Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __UpperCAmelCase = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __lowercase : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : List[Any] ,A : int ,A : Optional[Any]=13 ,A : Dict=7 ,A : Dict=True ,A : Any=False ,A : Dict=99 ,A : int=32 ,A : Optional[int]=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : str=0.1 ,A : int=0.1 ,A : Optional[int]=20 ,A : Tuple=2 ,A : str=1 ,A : Optional[Any]=0 ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : int = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Union[str, Any] = pad_token_id UpperCAmelCase__ : List[str] = bos_token_id def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] ,self.vocab_size ).clip(3 ,self.vocab_size ) UpperCAmelCase__ : List[str] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) ,1 ) UpperCAmelCase__ : Any = np.concatenate([input_ids, eos_tensor] ,axis=1 ) UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : str = self.config_cls( vocab_size=self.vocab_size ,d_model=self.hidden_size ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,eos_token_ids=[2] ,bos_token_id=self.bos_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.pad_token_id ,**self.config_updates ,) UpperCAmelCase__ : Optional[Any] = prepare_pegasus_inputs_dict(A ,A ,A ) return config, inputs_dict def __lowercase ( self : Any ,A : Optional[int] ,A : str ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : Dict = model_class_name(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) ,dtype="""i4""" ) UpperCAmelCase__ : str = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Optional[int] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : int = model.decode( decoder_input_ids[:, -1:] ,A ,decoder_attention_mask=A ,past_key_values=outputs_cache.past_key_values ,decoder_position_ids=A ,) UpperCAmelCase__ : Dict = model.decode(A ,A ) UpperCAmelCase__ : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def __lowercase ( self : Optional[int] ,A : str ,A : Dict ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : str = model_class_name(A ) UpperCAmelCase__ : Any = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] ,axis=-1 ,) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : Dict = model.decode( decoder_input_ids[:, -1:] ,A ,past_key_values=outputs_cache.past_key_values ,decoder_attention_mask=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Union[str, Any] = model.decode(A ,A ,decoder_attention_mask=A ) UpperCAmelCase__ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , ): '''simple docstring''' if attention_mask is None: UpperCAmelCase__ : Union[str, Any] = np.not_equal(__UpperCamelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Tuple = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) snake_case_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = FlaxPegasusModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ) def __lowercase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(A ,A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(A ,A ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : List[Any] = self._prepare_for_class(A ,A ) UpperCAmelCase__ : int = model_class(A ) @jax.jit def encode_jitted(A : Optional[int] ,A : Union[str, Any]=None ,**A : Optional[Any] ): return model.encode(input_ids=A ,attention_mask=A ) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : int = encode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : Dict = encode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : Dict = model_class(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ,inputs_dict["""attention_mask"""] ) UpperCAmelCase__ : Dict = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(A : List[Any] ,A : Any ,A : List[Any] ): return model.decode( decoder_input_ids=A ,decoder_attention_mask=A ,encoder_outputs=A ,) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : Tuple = decode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : str = decode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""google/pegasus-large""" ,from_pt=A ) UpperCAmelCase__ : Any = np.ones((1, 1) ) UpperCAmelCase__ : Optional[Any] = model(A ) self.assertIsNotNone(A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Optional[Any] = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Union[str, Any] = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] UpperCAmelCase__ : str = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] UpperCAmelCase__ : str = tokenizer(A ,return_tensors="""np""" ,truncation=A ,max_length=512 ,padding=A ) UpperCAmelCase__ : Union[str, Any] = model.generate(**A ,num_beams=2 ).sequences UpperCAmelCase__ : int = tokenizer.batch_decode(A ,skip_special_tokens=A ) assert tgt_text == decoded
65
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase ( __UpperCamelCase = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCAmelCase__ : Dict = F"{olid} is not a valid Open Library olid" raise ValueError(__UpperCamelCase ) return requests.get(F"https://openlibrary.org/{new_olid}.json" ).json() def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } UpperCAmelCase__ : Dict = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase__ : str = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] UpperCAmelCase__ : Dict = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Dict = """, """.join(__UpperCamelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __UpperCAmelCase = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: __UpperCAmelCase = summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('\n'.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
65
1
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""" , """False""" ) ) is not True , reason="""Skipping test because should only be run when releasing minor transformers version""" , ) @pytest.mark.usefixtures("""sm_env""" ) @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue_model_parallelism.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1_6_0_0, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1_6_0_0, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, ] ) class __lowercase ( unittest.TestCase ): def __lowercase ( self : str ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( f"cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py".split() ,encoding="""utf-8""" ,check=A ,) assert hasattr(self ,"""env""" ) def __lowercase ( self : Any ,A : Any ): '''simple docstring''' # configuration for running training on smdistributed Model Parallel UpperCAmelCase__ : Tuple = { """enabled""": True, """processes_per_host""": 8, } UpperCAmelCase__ : str = { """enabled""": True, """parameters""": { """microbatches""": 4, """placement_strategy""": """spread""", """pipeline""": """interleaved""", """optimize""": """speed""", """partitions""": 4, """ddp""": True, }, } UpperCAmelCase__ : Optional[int] = {"""smdistributed""": {"""modelparallel""": smp_options}, """mpi""": mpi_options} UpperCAmelCase__ : Union[str, Any] = """trainer""" if self.script == """run_glue.py""" else """smtrainer""" # creates estimator return HuggingFace( entry_point=self.script ,source_dir=self.env.test_path ,role=self.env.role ,image_uri=self.env.image_uri ,base_job_name=f"{self.env.base_job_name}-{instance_count}-smp-{name_extension}" ,instance_count=A ,instance_type=self.instance_type ,debugger_hook_config=A ,hyperparameters={ **self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path, """max_steps""": 500, } ,metric_definitions=self.env.metric_definitions ,distribution=A ,py_version="""py36""" ,) def __lowercase ( self : Dict ,A : Tuple ): '''simple docstring''' TrainingJobAnalytics(A ).export_csv(f"{self.env.test_path}/{job_name}_metrics.csv" ) @parameterized.expand([(1,)] ) def __lowercase ( self : Tuple ,A : Union[str, Any] ): '''simple docstring''' # create estimator UpperCAmelCase__ : Optional[int] = self.create_estimator(A ) # run training estimator.fit() # result dataframe UpperCAmelCase__ : List[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) UpperCAmelCase__ : List[str] = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ : Any = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" ,999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"{estimator.latest_training_job.name}.json" ,"""w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} ,A )
65
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=True , __UpperCamelCase="pt" ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {"""add_prefix_space""": True} if isinstance(__UpperCamelCase , __UpperCamelCase ) and not line.startswith(""" """ ) else {} UpperCAmelCase__ : List[str] = padding_side return tokenizer( [line] , max_length=__UpperCamelCase , padding="""max_length""" if pad_to_max_length else None , truncation=__UpperCamelCase , return_tensors=__UpperCamelCase , add_special_tokens=__UpperCamelCase , **__UpperCamelCase , ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , ): '''simple docstring''' UpperCAmelCase__ : str = input_ids.ne(__UpperCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ,A : List[Any] ,A : Union[str, Any] ,A : Any ,A : Optional[int] ,A : Union[str, Any]="train" ,A : Tuple=None ,A : Union[str, Any]=None ,A : Tuple=None ,A : int="" ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : Optional[Any] = Path(A ).joinpath(type_path + """.source""" ) UpperCAmelCase__ : List[str] = Path(A ).joinpath(type_path + """.target""" ) UpperCAmelCase__ : Dict = self.get_char_lens(self.src_file ) UpperCAmelCase__ : int = max_source_length UpperCAmelCase__ : List[str] = max_target_length assert min(self.src_lens ) > 0, f"found empty line in {self.src_file}" UpperCAmelCase__ : Dict = tokenizer UpperCAmelCase__ : str = prefix if n_obs is not None: UpperCAmelCase__ : int = self.src_lens[:n_obs] UpperCAmelCase__ : Any = src_lang UpperCAmelCase__ : Any = tgt_lang def __len__( self : Optional[Any] ): '''simple docstring''' return len(self.src_lens ) def __getitem__( self : Union[str, Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = index + 1 # linecache starts at 1 UpperCAmelCase__ : Tuple = self.prefix + linecache.getline(str(self.src_file ) ,A ).rstrip("""\n""" ) UpperCAmelCase__ : Dict = linecache.getline(str(self.tgt_file ) ,A ).rstrip("""\n""" ) assert source_line, f"empty source line for index {index}" assert tgt_line, f"empty tgt line for index {index}" # Need to add eos token manually for T5 if isinstance(self.tokenizer ,A ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right UpperCAmelCase__ : str = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,A ) else self.tokenizer ) UpperCAmelCase__ : Tuple = self.tokenizer.generator if isinstance(self.tokenizer ,A ) else self.tokenizer UpperCAmelCase__ : Tuple = encode_line(A ,A ,self.max_source_length ,"""right""" ) UpperCAmelCase__ : Dict = encode_line(A ,A ,self.max_target_length ,"""right""" ) UpperCAmelCase__ : Optional[Any] = source_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : List[str] = target_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : Union[str, Any] = source_inputs["""attention_mask"""].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def __lowercase ( A : int ): '''simple docstring''' return [len(A ) for x in Path(A ).open().readlines()] def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : int = torch.stack([x["""input_ids"""] for x in batch] ) UpperCAmelCase__ : Union[str, Any] = torch.stack([x["""attention_mask"""] for x in batch] ) UpperCAmelCase__ : Any = torch.stack([x["""decoder_input_ids"""] for x in batch] ) UpperCAmelCase__ : List[Any] = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : Any = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : str = trim_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = trim_batch(A ,A ,attention_mask=A ) UpperCAmelCase__ : List[str] = { """input_ids""": source_ids, """attention_mask""": source_mask, """decoder_input_ids""": y, } return batch __UpperCAmelCase = getLogger(__name__) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return list(itertools.chain.from_iterable(__UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = get_git_info() save_json(__UpperCamelCase , os.path.join(__UpperCamelCase , """git_log.json""" ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=4 , **__UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=__UpperCamelCase , **__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = git.Repo(search_parent_directories=__UpperCamelCase ) UpperCAmelCase__ : List[str] = { """repo_id""": str(__UpperCamelCase ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), """hostname""": str(socket.gethostname() ), } return repo_infos def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return list(map(__UpperCamelCase , __UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """wb""" ) as f: return pickle.dump(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' def remove_articles(__UpperCamelCase ): return re.sub(r"""\b(a|an|the)\b""" , """ """ , __UpperCamelCase ) def white_space_fix(__UpperCamelCase ): return " ".join(text.split() ) def remove_punc(__UpperCamelCase ): UpperCAmelCase__ : List[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCamelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCamelCase ) ) ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : Dict = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : int = Counter(__UpperCamelCase ) & Counter(__UpperCamelCase ) UpperCAmelCase__ : List[str] = sum(common.values() ) if num_same == 0: return 0 UpperCAmelCase__ : str = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Tuple = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return normalize_answer(__UpperCamelCase ) == normalize_answer(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' assert len(__UpperCamelCase ) == len(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 0 for hypo, pred in zip(__UpperCamelCase , __UpperCamelCase ): em += exact_match_score(__UpperCamelCase , __UpperCamelCase ) if len(__UpperCamelCase ) > 0: em /= len(__UpperCamelCase ) return {"em": em} def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return model_prefix.startswith("""rag""" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead UpperCAmelCase__ : str = """dropout_rate""" for p in extra_params: if getattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): if not hasattr(__UpperCamelCase , __UpperCamelCase ) and not hasattr(__UpperCamelCase , equivalent_param[p] ): logger.info("""config doesn't have a `{}` attribute""".format(__UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) continue UpperCAmelCase__ : Tuple = p if hasattr(__UpperCamelCase , __UpperCamelCase ) else equivalent_param[p] setattr(__UpperCamelCase , __UpperCamelCase , getattr(__UpperCamelCase , __UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) return hparams, config
65
1
"""simple docstring""" import os import re import shutil import sys import tempfile import unittest import black __UpperCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If DDPMSchedulerOutput is changed in scheduling_ddpm.py, this code needs to be manually updated. __UpperCAmelCase = ' \"""\n Output class for the scheduler\'s step function output.\n\n Args:\n prev_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images):\n Computed sample (x_{t-1}) of previous timestep. `prev_sample` should be used as next model input in the\n denoising loop.\n pred_original_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images):\n The predicted denoised sample (x_{0}) based on the model output from the current timestep.\n `pred_original_sample` can be used to preview progress or for guidance.\n \"""\n\n prev_sample: torch.FloatTensor\n pred_original_sample: Optional[torch.FloatTensor] = None\n' class __lowercase ( unittest.TestCase ): def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = tempfile.mkdtemp() os.makedirs(os.path.join(self.diffusers_dir ,"""schedulers/""" ) ) UpperCAmelCase__ : Optional[Any] = self.diffusers_dir shutil.copy( os.path.join(A ,"""src/diffusers/schedulers/scheduling_ddpm.py""" ) ,os.path.join(self.diffusers_dir ,"""schedulers/scheduling_ddpm.py""" ) ,) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = """src/diffusers""" shutil.rmtree(self.diffusers_dir ) def __lowercase ( self : List[str] ,A : Dict ,A : List[Any] ,A : int ,A : Tuple=None ): '''simple docstring''' UpperCAmelCase__ : Dict = comment + f"\nclass {class_name}(nn.Module):\n" + class_code if overwrite_result is not None: UpperCAmelCase__ : str = comment + f"\nclass {class_name}(nn.Module):\n" + overwrite_result UpperCAmelCase__ : Tuple = black.Mode(target_versions={black.TargetVersion.PYaa} ,line_length=119 ) UpperCAmelCase__ : List[str] = black.format_str(A ,mode=A ) UpperCAmelCase__ : List[str] = os.path.join(self.diffusers_dir ,"""new_code.py""" ) with open(A ,"""w""" ,newline="""\n""" ) as f: f.write(A ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(A ) ) == 0 ) else: check_copies.is_copy_consistent(f.name ,overwrite=A ) with open(A ,"""r""" ) as f: self.assertTrue(f.read() ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Dict = check_copies.find_code_in_diffusers("""schedulers.scheduling_ddpm.DDPMSchedulerOutput""" ) self.assertEqual(A ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' # Base copy consistency self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput""" ,"""DDPMSchedulerOutput""" ,REFERENCE_CODE + """\n""" ,) # With no empty line at the end self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput""" ,"""DDPMSchedulerOutput""" ,A ,) # Copy consistency with rename self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test""" ,"""TestSchedulerOutput""" ,re.sub("""DDPM""" ,"""Test""" ,A ) ,) # Copy consistency with a really long name UpperCAmelCase__ : Optional[Any] = """TestClassWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason""" self.check_copy_consistency( f"# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->{long_class_name}" ,f"{long_class_name}SchedulerOutput" ,re.sub("""Bert""" ,A ,A ) ,) # Copy consistency with overwrite self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test""" ,"""TestSchedulerOutput""" ,A ,overwrite_result=re.sub("""DDPM""" ,"""Test""" ,A ) ,)
65
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaControlnetPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return 32 @property def __lowercase ( self : int ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self : Any ): '''simple docstring''' return 100 @property def __lowercase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Tuple = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } UpperCAmelCase__ : int = UNetaDConditionModel(**A ) return model @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : str = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : str = self.dummy_unet UpperCAmelCase__ : List[Any] = self.dummy_movq UpperCAmelCase__ : List[Any] = DDIMScheduler( num_train_timesteps=1_000 ,beta_schedule="""linear""" ,beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,clip_sample=A ,set_alpha_to_one=A ,steps_offset=1 ,prediction_type="""epsilon""" ,thresholding=A ,) UpperCAmelCase__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __lowercase ( self : str ,A : Optional[Any] ,A : Any=0 ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( A ) # create hint UpperCAmelCase__ : int = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) if str(A ).startswith("""mps""" ): UpperCAmelCase__ : Optional[int] = torch.manual_seed(A ) else: UpperCAmelCase__ : Dict = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase__ : Dict = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = """cpu""" UpperCAmelCase__ : List[Any] = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = self.pipeline_class(**A ) UpperCAmelCase__ : Optional[int] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = pipe(**self.get_dummy_inputs(A ) ) UpperCAmelCase__ : Tuple = output.images UpperCAmelCase__ : Dict = pipe( **self.get_dummy_inputs(A ) ,return_dict=A ,)[0] UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1] UpperCAmelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Optional[int] = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : Union[str, Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) UpperCAmelCase__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) UpperCAmelCase__ : int = torch.from_numpy(np.array(A ) ).float() / 2_5_5.0 UpperCAmelCase__ : Union[str, Any] = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) UpperCAmelCase__ : List[str] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" ,torch_dtype=torch.floataa ) pipe_prior.to(A ) UpperCAmelCase__ : List[Any] = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" ,torch_dtype=torch.floataa ) UpperCAmelCase__ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[Any] = """A robot, 4k photo""" UpperCAmelCase__ : List[Any] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = pipe_prior( A ,generator=A ,num_inference_steps=5 ,negative_prompt="""""" ,).to_tuple() UpperCAmelCase__ : List[str] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ : int = pipeline( image_embeds=A ,negative_image_embeds=A ,hint=A ,generator=A ,num_inference_steps=100 ,output_type="""np""" ,) UpperCAmelCase__ : Any = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A ,A )
65
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'shi-labs/dinat-mini-in1k-224': 'https://huggingface.co/shi-labs/dinat-mini-in1k-224/resolve/main/config.json', # See all Dinat models at https://huggingface.co/models?filter=dinat } class __lowercase ( __lowerCamelCase , __lowerCamelCase ): snake_case_ = """dinat""" snake_case_ = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : Union[str, Any] ,A : Tuple=4 ,A : Dict=3 ,A : Union[str, Any]=64 ,A : List[str]=[3, 4, 6, 5] ,A : Any=[2, 4, 8, 16] ,A : Optional[Any]=7 ,A : Optional[Any]=[[1, 8, 1], [1, 4, 1, 4], [1, 2, 1, 2, 1, 2], [1, 1, 1, 1, 1]] ,A : List[Any]=3.0 ,A : int=True ,A : int=0.0 ,A : List[str]=0.0 ,A : Any=0.1 ,A : List[str]="gelu" ,A : Dict=0.0_2 ,A : Optional[int]=1e-5 ,A : Dict=0.0 ,A : str=None ,A : List[Any]=None ,**A : int ,): '''simple docstring''' super().__init__(**A ) UpperCAmelCase__ : str = patch_size UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : Optional[int] = embed_dim UpperCAmelCase__ : int = depths UpperCAmelCase__ : str = len(A ) UpperCAmelCase__ : List[str] = num_heads UpperCAmelCase__ : Optional[Any] = kernel_size UpperCAmelCase__ : str = dilations UpperCAmelCase__ : Union[str, Any] = mlp_ratio UpperCAmelCase__ : Optional[Any] = qkv_bias UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = drop_path_rate UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Optional[int] = layer_norm_eps UpperCAmelCase__ : str = initializer_range # we set the hidden_size attribute in order to make Dinat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCAmelCase__ : Optional[int] = int(embed_dim * 2 ** (len(A ) - 1) ) UpperCAmelCase__ : Optional[Any] = layer_scale_init_value UpperCAmelCase__ : Optional[Any] = ["""stem"""] + [f"stage{idx}" for idx in range(1 ,len(A ) + 1 )] UpperCAmelCase__ , UpperCAmelCase__ : int = get_aligned_output_features_output_indices( out_features=A ,out_indices=A ,stage_names=self.stage_names )
65
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = """vision-encoder-decoder""" snake_case_ = True def __init__( self : List[Any] ,**A : Union[str, Any] ): '''simple docstring''' super().__init__(**A ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) UpperCAmelCase__ : int = kwargs.pop("""encoder""" ) UpperCAmelCase__ : int = encoder_config.pop("""model_type""" ) UpperCAmelCase__ : str = kwargs.pop("""decoder""" ) UpperCAmelCase__ : Dict = decoder_config.pop("""model_type""" ) UpperCAmelCase__ : List[Any] = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Any = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Union[str, Any] = True @classmethod def __lowercase ( cls : List[Any] ,A : PretrainedConfig ,A : PretrainedConfig ,**A : Tuple ): '''simple docstring''' logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : List[Any] = True return cls(encoder=encoder_config.to_dict() ,decoder=decoder_config.to_dict() ,**A ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Dict = self.encoder.to_dict() UpperCAmelCase__ : Any = self.decoder.to_dict() UpperCAmelCase__ : Dict = self.__class__.model_type return output class __lowercase ( __lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' return 1e-4 @property def __lowercase ( self : List[Any] ): '''simple docstring''' return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : Optional[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : List[str] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def __lowercase ( self : Dict ,A : "PreTrainedTokenizerBase" ,A : int = -1 ,A : int = -1 ,A : bool = False ,A : Optional["TensorType"] = None ,): '''simple docstring''' import torch UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : List[Any] = super().generate_dummy_inputs( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) UpperCAmelCase__ , UpperCAmelCase__ : int = dummy_input["""input_ids"""].shape UpperCAmelCase__ : int = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCAmelCase__ : Tuple = dummy_input.pop("""input_ids""" ) UpperCAmelCase__ : Optional[int] = dummy_input.pop("""attention_mask""" ) UpperCAmelCase__ : Dict = torch.zeros(A ) return common_inputs class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : Any ,A : PretrainedConfig ): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(A ) def __lowercase ( self : Dict ,A : PretrainedConfig ,A : PretrainedConfig ,A : str = "default" ): '''simple docstring''' UpperCAmelCase__ : List[str] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(A ,A )
65
1
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __lowercase ( __lowerCamelCase ): def __init__( self : Optional[Any] ,A : str ,A : Tuple ): '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM UpperCAmelCase__ : Any = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=A ,scheduler=A ) @torch.no_grad() def __call__( self : List[Any] ,A : int = 1 ,A : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,A : float = 0.0 ,A : int = 50 ,A : Optional[bool] = None ,A : Optional[str] = "pil" ,A : bool = True ,): '''simple docstring''' # Sample gaussian noise to begin loop if isinstance(self.unet.config.sample_size ,A ): UpperCAmelCase__ : Optional[int] = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: UpperCAmelCase__ : str = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(A ,A ) and len(A ) != batch_size: raise ValueError( f"You have passed a list of generators of length {len(A )}, but requested an effective batch" f" size of {batch_size}. Make sure the batch size matches the length of the generators." ) UpperCAmelCase__ : Optional[Any] = randn_tensor(A ,generator=A ,device=self.device ,dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(A ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output UpperCAmelCase__ : Tuple = self.unet(A ,A ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 UpperCAmelCase__ : Tuple = self.scheduler.step( A ,A ,A ,eta=A ,use_clipped_model_output=A ,generator=A ).prev_sample UpperCAmelCase__ : Optional[Any] = (image / 2 + 0.5).clamp(0 ,1 ) UpperCAmelCase__ : Optional[int] = image.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": UpperCAmelCase__ : List[str] = self.numpy_to_pil(A ) if not return_dict: return (image,) return ImagePipelineOutput(images=A )
65
"""simple docstring""" import requests def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = {"""Content-Type""": """application/json"""} UpperCAmelCase__ : Optional[Any] = requests.post(__UpperCamelCase , json={"""text""": message_body} , headers=__UpperCamelCase ) if response.status_code != 200: UpperCAmelCase__ : Any = ( """Request to slack returned an error """ F"{response.status_code}, the response is:\n{response.text}" ) raise ValueError(__UpperCamelCase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
65
1
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = CTRLTokenizer snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ : List[Any] = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase__ : Optional[int] = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : List[Any] = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase__ : int = {"""unk_token""": """<unk>"""} UpperCAmelCase__ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(A ) ) def __lowercase ( self : int ,**A : Dict ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """adapt react readapt apt""" UpperCAmelCase__ : Any = """adapt react readapt apt""" return input_text, output_text def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = CTRLTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) UpperCAmelCase__ : Tuple = """adapt react readapt apt""" UpperCAmelCase__ : Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase__ : Dict = tokenizer.tokenize(A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Any = tokens + [tokenizer.unk_token] UpperCAmelCase__ : Dict = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,A )
65
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = CTRLTokenizer snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ : List[Any] = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase__ : Optional[int] = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : List[Any] = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase__ : int = {"""unk_token""": """<unk>"""} UpperCAmelCase__ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(A ) ) def __lowercase ( self : int ,**A : Dict ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """adapt react readapt apt""" UpperCAmelCase__ : Any = """adapt react readapt apt""" return input_text, output_text def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = CTRLTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) UpperCAmelCase__ : Tuple = """adapt react readapt apt""" UpperCAmelCase__ : Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase__ : Dict = tokenizer.tokenize(A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Any = tokens + [tokenizer.unk_token] UpperCAmelCase__ : Dict = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,A )
65
1
"""simple docstring""" from abc import ABC, abstractmethod from typing import List, Optional class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ): '''simple docstring''' # test for the above condition self.test() def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = 0 UpperCAmelCase__ : List[str] = False while not completed: if counter == 1: self.reset() UpperCAmelCase__ : List[str] = self.advance() if not self.does_advance(A ): raise Exception( """Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.""" ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = self.update(A ) counter += 1 if counter > 10_000: raise Exception("""update() does not fulfill the constraint.""" ) if self.remaining() != 0: raise Exception("""Custom Constraint is not defined correctly.""" ) @abstractmethod def __lowercase ( self : Dict ): '''simple docstring''' raise NotImplementedError( f"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def __lowercase ( self : str ,A : int ): '''simple docstring''' raise NotImplementedError( f"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def __lowercase ( self : Optional[Any] ,A : int ): '''simple docstring''' raise NotImplementedError( f"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def __lowercase ( self : Any ): '''simple docstring''' raise NotImplementedError( f"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def __lowercase ( self : Any ): '''simple docstring''' raise NotImplementedError( f"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) @abstractmethod def __lowercase ( self : str ,A : Optional[int]=False ): '''simple docstring''' raise NotImplementedError( f"{self.__class__} is an abstract class. Only classes inheriting this class can be called." ) class __lowercase ( __lowerCamelCase ): def __init__( self : List[str] ,A : List[int] ): '''simple docstring''' super(A ,self ).__init__() if not isinstance(A ,A ) or len(A ) == 0: raise ValueError(f"`token_ids` has to be a non-empty list, but is {token_ids}." ) if any((not isinstance(A ,A ) or token_id < 0) for token_id in token_ids ): raise ValueError(f"Each list in `token_ids` has to be a list of positive integers, but is {token_ids}." ) UpperCAmelCase__ : Optional[int] = token_ids UpperCAmelCase__ : List[Any] = len(self.token_ids ) UpperCAmelCase__ : str = -1 # the index of the currently fulfilled step UpperCAmelCase__ : Union[str, Any] = False def __lowercase ( self : Dict ): '''simple docstring''' if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def __lowercase ( self : Optional[Any] ,A : int ): '''simple docstring''' if not isinstance(A ,A ): raise ValueError(f"`token_id` has to be an `int`, but is {token_id} of type {type(A )}" ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def __lowercase ( self : str ,A : int ): '''simple docstring''' if not isinstance(A ,A ): raise ValueError(f"`token_id` has to be an `int`, but is {token_id} of type {type(A )}" ) UpperCAmelCase__ : Union[str, Any] = False UpperCAmelCase__ : Dict = False UpperCAmelCase__ : Union[str, Any] = False if self.does_advance(A ): self.fulfilled_idx += 1 UpperCAmelCase__ : Tuple = True if self.fulfilled_idx == (self.seqlen - 1): UpperCAmelCase__ : str = True UpperCAmelCase__ : int = completed else: # failed to make progress. UpperCAmelCase__ : Optional[Any] = True self.reset() return stepped, completed, reset def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Optional[Any] = 0 def __lowercase ( self : Optional[int] ): '''simple docstring''' return self.seqlen - (self.fulfilled_idx + 1) def __lowercase ( self : List[str] ,A : List[str]=False ): '''simple docstring''' UpperCAmelCase__ : Any = PhrasalConstraint(self.token_ids ) if stateful: UpperCAmelCase__ : int = self.seqlen UpperCAmelCase__ : Optional[int] = self.fulfilled_idx UpperCAmelCase__ : int = self.completed return new_constraint class __lowercase : def __init__( self : Dict ,A : List[List[int]] ,A : Tuple=True ): '''simple docstring''' UpperCAmelCase__ : str = max([len(A ) for one in nested_token_ids] ) UpperCAmelCase__ : List[Any] = {} for token_ids in nested_token_ids: UpperCAmelCase__ : Tuple = root for tidx, token_id in enumerate(A ): if token_id not in level: UpperCAmelCase__ : Dict = {} UpperCAmelCase__ : Optional[Any] = level[token_id] if no_subsets and self.has_subsets(A ,A ): raise ValueError( """Each list in `nested_token_ids` can't be a complete subset of another list, but is""" f" {nested_token_ids}." ) UpperCAmelCase__ : int = root def __lowercase ( self : List[str] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.trie for current_token in current_seq: UpperCAmelCase__ : int = start[current_token] UpperCAmelCase__ : str = list(start.keys() ) return next_tokens def __lowercase ( self : List[str] ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : int = self.next_tokens(A ) return len(A ) == 0 def __lowercase ( self : Union[str, Any] ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = list(root.values() ) if len(A ) == 0: return 1 else: return sum([self.count_leaves(A ) for nn in next_nodes] ) def __lowercase ( self : Optional[Any] ,A : List[Any] ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = self.count_leaves(A ) return len(A ) != leaf_count class __lowercase ( __lowerCamelCase ): def __init__( self : str ,A : List[List[int]] ): '''simple docstring''' super(A ,self ).__init__() if not isinstance(A ,A ) or len(A ) == 0: raise ValueError(f"`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}." ) if any(not isinstance(A ,A ) for token_ids in nested_token_ids ): raise ValueError(f"`nested_token_ids` has to be a list of lists, but is {nested_token_ids}." ) if any( any((not isinstance(A ,A ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( f"Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}." ) UpperCAmelCase__ : List[Any] = DisjunctiveTrie(A ) UpperCAmelCase__ : Dict = nested_token_ids UpperCAmelCase__ : Optional[Any] = self.trie.max_height UpperCAmelCase__ : str = [] UpperCAmelCase__ : Union[str, Any] = False def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.trie.next_tokens(self.current_seq ) if len(A ) == 0: return None else: return token_list def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' if not isinstance(A ,A ): raise ValueError(f"`token_id` is supposed to be type `int`, but is {token_id} of type {type(A )}" ) UpperCAmelCase__ : List[Any] = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def __lowercase ( self : Dict ,A : int ): '''simple docstring''' if not isinstance(A ,A ): raise ValueError(f"`token_id` is supposed to be type `int`, but is {token_id} of type {type(A )}" ) UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : Optional[Any] = False if self.does_advance(A ): self.current_seq.append(A ) UpperCAmelCase__ : Dict = True else: UpperCAmelCase__ : Optional[int] = True self.reset() UpperCAmelCase__ : Optional[Any] = self.trie.reached_leaf(self.current_seq ) UpperCAmelCase__ : int = completed return stepped, completed, reset def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : str = False UpperCAmelCase__ : List[str] = [] def __lowercase ( self : int ): '''simple docstring''' if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def __lowercase ( self : Optional[int] ,A : Optional[int]=False ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = DisjunctiveConstraint(self.token_ids ) if stateful: UpperCAmelCase__ : Optional[int] = self.seqlen UpperCAmelCase__ : Tuple = self.current_seq UpperCAmelCase__ : List[str] = self.completed return new_constraint class __lowercase : def __init__( self : List[str] ,A : List[Constraint] ): '''simple docstring''' UpperCAmelCase__ : str = constraints # max # of steps required to fulfill a given constraint UpperCAmelCase__ : List[str] = max([c.seqlen for c in constraints] ) UpperCAmelCase__ : Any = len(A ) UpperCAmelCase__ : List[Any] = False self.init_state() def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = [] UpperCAmelCase__ : Optional[Any] = None UpperCAmelCase__ : Dict = [constraint.copy(stateful=A ) for constraint in self.constraints] def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Tuple = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" UpperCAmelCase__ : str = constraint.advance() if isinstance(A ,A ): token_list.append(A ) elif isinstance(A ,A ): token_list.extend(A ) else: UpperCAmelCase__ : List[Any] = self.inprogress_constraint.advance() if isinstance(A ,A ): token_list.append(A ) elif isinstance(A ,A ): token_list.extend(A ) if len(A ) == 0: return None else: return token_list def __lowercase ( self : Dict ,A : Optional[List[int]] ): '''simple docstring''' self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = self.add(A ) # the entire list of constraints are fulfilled if self.completed: break def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' if not isinstance(A ,A ): raise ValueError(f"`token_id` should be an `int`, but is `{token_id}`." ) UpperCAmelCase__ , UpperCAmelCase__ : int = False, False if self.completed: UpperCAmelCase__ : Dict = True UpperCAmelCase__ : Optional[Any] = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.inprogress_constraint.update(A ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=A ) ) UpperCAmelCase__ : Optional[int] = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) UpperCAmelCase__ : Optional[Any] = None if len(self.pending_constraints ) == 0: # we're done! UpperCAmelCase__ : Optional[Any] = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(A ): UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : str = pending_constraint.update(A ) if not stepped: raise Exception( """`constraint.update(token_id)` is not yielding incremental progress, """ """even though `constraint.does_advance(token_id)` is true.""" ) if complete: self.complete_constraints.append(A ) UpperCAmelCase__ : str = None if not complete and stepped: UpperCAmelCase__ : List[str] = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". UpperCAmelCase__ : Any = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. UpperCAmelCase__ : Tuple = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def __lowercase ( self : Any ,A : Optional[int]=True ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: UpperCAmelCase__ : List[str] = [ constraint.copy(stateful=A ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: UpperCAmelCase__ : Optional[Any] = self.inprogress_constraint.copy(stateful=A ) UpperCAmelCase__ : List[Any] = [constraint.copy() for constraint in self.pending_constraints] return new_state
65
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCAmelCase = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfig', 'BridgeTowerTextConfig', 'BridgeTowerVisionConfig', ], 'processing_bridgetower': ['BridgeTowerProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['BridgeTowerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST', 'BridgeTowerForContrastiveLearning', 'BridgeTowerForImageAndTextRetrieval', 'BridgeTowerForMaskedLM', 'BridgeTowerModel', 'BridgeTowerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
65
1
"""simple docstring""" class __lowercase : def __init__( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = {} def __lowercase ( self : Any ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(A ,""" -> """ ,""" -> """.join([str(A ) for j in self.vertex[i]] ) ) def __lowercase ( self : Tuple ,A : int ,A : int ): '''simple docstring''' # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(A ) else: # else make a new vertex UpperCAmelCase__ : Optional[int] = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' # visited array for storing already visited nodes UpperCAmelCase__ : List[Any] = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(A ,A ) def __lowercase ( self : List[Any] ,A : int ,A : list ): '''simple docstring''' # mark start vertex as visited UpperCAmelCase__ : List[Any] = True print(A ,end=""" """ ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(A ,A ) if __name__ == "__main__": __UpperCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('DFS:') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
65
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """is_longer"""] def __init__( self : str ,A : Union[str, Any]=64 ,A : Tuple=48_000 ,A : Dict=480 ,A : List[str]=10 ,A : str=1_024 ,A : Any=0.0 ,A : Optional[int]=False ,A : float = 0 ,A : float = 14_000 ,A : int = None ,A : str = "fusion" ,A : str = "repeatpad" ,**A : List[Any] ,): '''simple docstring''' super().__init__( feature_size=A ,sampling_rate=A ,padding_value=A ,return_attention_mask=A ,**A ,) UpperCAmelCase__ : List[Any] = top_db UpperCAmelCase__ : Union[str, Any] = truncation UpperCAmelCase__ : Optional[int] = padding UpperCAmelCase__ : List[Any] = fft_window_size UpperCAmelCase__ : Optional[Any] = (fft_window_size >> 1) + 1 UpperCAmelCase__ : Any = hop_length UpperCAmelCase__ : List[str] = max_length_s UpperCAmelCase__ : List[Any] = max_length_s * sampling_rate UpperCAmelCase__ : List[Any] = sampling_rate UpperCAmelCase__ : Optional[int] = frequency_min UpperCAmelCase__ : Tuple = frequency_max UpperCAmelCase__ : List[str] = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm=A ,mel_scale="""htk""" ,) UpperCAmelCase__ : str = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm="""slaney""" ,mel_scale="""slaney""" ,) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def __lowercase ( self : List[str] ,A : np.array ,A : Optional[np.array] = None ): '''simple docstring''' UpperCAmelCase__ : Dict = spectrogram( A ,window_function(self.fft_window_size ,"""hann""" ) ,frame_length=self.fft_window_size ,hop_length=self.hop_length ,power=2.0 ,mel_filters=A ,log_mel="""dB""" ,) return log_mel_spectrogram.T def __lowercase ( self : Optional[Any] ,A : Union[str, Any] ,A : int ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = np.array_split(list(range(0 ,total_frames - chunk_frames + 1 ) ) ,3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : List[str] = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : int = [0] # randomly choose index for each part UpperCAmelCase__ : Tuple = np.random.choice(ranges[0] ) UpperCAmelCase__ : Tuple = np.random.choice(ranges[1] ) UpperCAmelCase__ : str = np.random.choice(ranges[2] ) UpperCAmelCase__ : List[str] = mel[idx_front : idx_front + chunk_frames, :] UpperCAmelCase__ : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] UpperCAmelCase__ : Dict = mel[idx_back : idx_back + chunk_frames, :] UpperCAmelCase__ : Optional[Any] = torch.tensor(mel[None, None, :] ) UpperCAmelCase__ : int = torch.nn.functional.interpolate( A ,size=[chunk_frames, 64] ,mode="""bilinear""" ,align_corners=A ) UpperCAmelCase__ : Dict = mel_shrink[0][0].numpy() UpperCAmelCase__ : Dict = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] ,axis=0 ) return mel_fusion def __lowercase ( self : Any ,A : np.array ,A : Optional[int] ,A : Any ,A : Tuple ): '''simple docstring''' if waveform.shape[0] > max_length: if truncation == "rand_trunc": UpperCAmelCase__ : int = True # random crop to max_length (for compatibility) -> this should be handled by self.pad UpperCAmelCase__ : str = len(A ) - max_length UpperCAmelCase__ : Optional[Any] = np.random.randint(0 ,overflow + 1 ) UpperCAmelCase__ : Optional[int] = waveform[idx : idx + max_length] UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] elif truncation == "fusion": UpperCAmelCase__ : Tuple = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : Optional[int] = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed UpperCAmelCase__ : int = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. UpperCAmelCase__ : List[Any] = np.stack([mel, mel, mel, mel] ,axis=0 ) UpperCAmelCase__ : Any = False else: UpperCAmelCase__ : Union[str, Any] = self._random_mel_fusion(A ,A ,A ) UpperCAmelCase__ : List[str] = True else: raise NotImplementedError(f"data_truncating {truncation} not implemented" ) else: UpperCAmelCase__ : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": UpperCAmelCase__ : str = int(max_length / len(A ) ) UpperCAmelCase__ : int = np.stack(np.tile(A ,n_repeat + 1 ) )[:max_length] if padding == "repeatpad": UpperCAmelCase__ : List[Any] = int(max_length / len(A ) ) UpperCAmelCase__ : str = np.stack(np.tile(A ,A ) ) UpperCAmelCase__ : Optional[Any] = np.pad(A ,(0, max_length - waveform.shape[0]) ,mode="""constant""" ,constant_values=0 ) if truncation == "fusion": UpperCAmelCase__ : int = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] ,axis=0 ) else: UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self : str ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : str = None ,A : Optional[str] = None ,A : Optional[int] = None ,A : Optional[int] = None ,A : Optional[Union[str, TensorType]] = None ,**A : List[str] ,): '''simple docstring''' UpperCAmelCase__ : Optional[int] = truncation if truncation is not None else self.truncation UpperCAmelCase__ : Dict = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" f" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" f" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[int] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : List[str] = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Any = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [np.asarray(A )] # convert to mel spectrogram, truncate and pad if needed. UpperCAmelCase__ : Tuple = [ self._get_input_mel(A ,max_length if max_length else self.nb_max_samples ,A ,A ) for waveform in raw_speech ] UpperCAmelCase__ : Optional[int] = [] UpperCAmelCase__ : Tuple = [] for mel, longer in padded_inputs: input_mel.append(A ) is_longer.append(A ) if truncation == "fusion" and sum(A ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer UpperCAmelCase__ : List[str] = np.random.randint(0 ,len(A ) ) UpperCAmelCase__ : int = True if isinstance(input_mel[0] ,A ): UpperCAmelCase__ : Tuple = [np.asarray(A ,dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool UpperCAmelCase__ : List[str] = [[longer] for longer in is_longer] UpperCAmelCase__ : List[Any] = {"""input_features""": input_mel, """is_longer""": is_longer} UpperCAmelCase__ : str = BatchFeature(A ) if return_tensors is not None: UpperCAmelCase__ : int = input_features.convert_to_tensors(A ) return input_features
65
1
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __lowercase : def __init__( self : Optional[Any] ,A : Any ,A : List[str]=2 ,A : Optional[Any]=8 ,A : List[Any]=True ,A : str=True ,A : Dict=True ,A : Dict=True ,A : Any=99 ,A : Dict=16 ,A : Optional[int]=5 ,A : Tuple=2 ,A : str=36 ,A : List[Any]="gelu" ,A : str=0.0 ,A : str=0.0 ,A : List[str]=512 ,A : Dict=16 ,A : int=2 ,A : int=0.0_2 ,A : List[Any]=3 ,A : Dict=4 ,A : int=None ,): '''simple docstring''' UpperCAmelCase__ : Any = parent UpperCAmelCase__ : Tuple = batch_size UpperCAmelCase__ : Tuple = seq_length UpperCAmelCase__ : List[Any] = is_training UpperCAmelCase__ : str = use_input_mask UpperCAmelCase__ : str = use_token_type_ids UpperCAmelCase__ : str = use_labels UpperCAmelCase__ : List[Any] = vocab_size UpperCAmelCase__ : Tuple = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : Optional[int] = num_attention_heads UpperCAmelCase__ : Union[str, Any] = intermediate_size UpperCAmelCase__ : List[str] = hidden_act UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Optional[Any] = type_vocab_size UpperCAmelCase__ : Optional[Any] = type_sequence_label_size UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Optional[int] = num_labels UpperCAmelCase__ : Tuple = num_choices UpperCAmelCase__ : List[Any] = scope def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : List[Any] = None if self.use_input_mask: UpperCAmelCase__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Optional[int] = None if self.use_token_type_ids: UpperCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase__ : Optional[Any] = None UpperCAmelCase__ : Tuple = None UpperCAmelCase__ : Optional[Any] = None if self.use_labels: UpperCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase__ : Tuple = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase__ : Any = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowercase ( self : str ): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A ,initializer_range=self.initializer_range ,) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.get_config() UpperCAmelCase__ : Dict = 300 return config def __lowercase ( self : Dict ): '''simple docstring''' ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : List[str] = self.prepare_config_and_inputs() UpperCAmelCase__ : Dict = True UpperCAmelCase__ : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowercase ( self : List[Any] ,A : List[Any] ,A : Any ,A : List[Any] ,A : List[str] ,A : int ,A : Dict ,A : str ): '''simple docstring''' UpperCAmelCase__ : List[str] = MraModel(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : str = model(A ,attention_mask=A ,token_type_ids=A ) UpperCAmelCase__ : Union[str, Any] = model(A ,token_type_ids=A ) UpperCAmelCase__ : Any = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def __lowercase ( self : Dict ,A : Union[str, Any] ,A : Any ,A : Any ,A : int ,A : Union[str, Any] ,A : List[str] ,A : int ,A : int ,A : Any ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : List[Any] = MraModel(A ) model.to(A ) model.eval() UpperCAmelCase__ : Optional[int] = model( A ,attention_mask=A ,token_type_ids=A ,encoder_hidden_states=A ,encoder_attention_mask=A ,) UpperCAmelCase__ : Optional[int] = model( A ,attention_mask=A ,token_type_ids=A ,encoder_hidden_states=A ,) UpperCAmelCase__ : int = model(A ,attention_mask=A ,token_type_ids=A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def __lowercase ( self : Union[str, Any] ,A : List[Any] ,A : Tuple ,A : Dict ,A : List[Any] ,A : Any ,A : Any ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Tuple = MraForMaskedLM(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : int = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def __lowercase ( self : Any ,A : int ,A : int ,A : Any ,A : List[Any] ,A : Any ,A : Tuple ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = MraForQuestionAnswering(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : Dict = model( A ,attention_mask=A ,token_type_ids=A ,start_positions=A ,end_positions=A ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def __lowercase ( self : List[Any] ,A : List[Any] ,A : int ,A : str ,A : Any ,A : int ,A : str ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.num_labels UpperCAmelCase__ : Any = MraForSequenceClassification(A ) model.to(A ) model.eval() UpperCAmelCase__ : Union[str, Any] = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def __lowercase ( self : List[str] ,A : Optional[int] ,A : List[Any] ,A : str ,A : Optional[int] ,A : int ,A : Union[str, Any] ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.num_labels UpperCAmelCase__ : Any = MraForTokenClassification(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : List[str] = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def __lowercase ( self : Optional[Any] ,A : Tuple ,A : Tuple ,A : Tuple ,A : int ,A : Optional[int] ,A : int ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.num_choices UpperCAmelCase__ : List[str] = MraForMultipleChoice(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : str = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase__ : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase__ : Any = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase__ : str = model( A ,attention_mask=A ,token_type_ids=A ,labels=A ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : List[Any] = config_and_inputs UpperCAmelCase__ : Any = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = () def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = MraModelTester(self ) UpperCAmelCase__ : Tuple = ConfigTester(self ,config_class=A ,hidden_size=37 ) def __lowercase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase__ : List[str] = type self.model_tester.create_and_check_model(*A ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) @slow def __lowercase ( self : int ): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : str = MraModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""MRA does not output attentions""" ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return @require_torch class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Any = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) UpperCAmelCase__ : str = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase__ : Dict = model(A )[0] UpperCAmelCase__ : Tuple = torch.Size((1, 256, 768) ) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Optional[int] = torch.tensor( [[[-0.0_1_4_0, 0.0_8_3_0, -0.0_3_8_1], [0.1_5_4_6, 0.1_4_0_2, 0.0_2_2_0], [0.1_1_6_2, 0.0_8_5_1, 0.0_1_6_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,A ,atol=1e-4 ) ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) UpperCAmelCase__ : Tuple = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase__ : Dict = model(A )[0] UpperCAmelCase__ : Union[str, Any] = 50_265 UpperCAmelCase__ : Optional[Any] = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : List[str] = torch.tensor( [[[9.2_5_9_5, -3.6_0_3_8, 1_1.8_8_1_9], [9.3_8_6_9, -3.2_6_9_3, 1_1.0_9_5_6], [1_1.8_5_2_4, -3.4_9_3_8, 1_3.1_2_1_0]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,A ,atol=1e-4 ) ) @slow def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) UpperCAmelCase__ : Optional[int] = torch.arange(4_096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase__ : Any = model(A )[0] UpperCAmelCase__ : Tuple = 50_265 UpperCAmelCase__ : Tuple = torch.Size((1, 4_096, vocab_size) ) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Tuple = torch.tensor( [[[5.4_7_8_9, -2.3_5_6_4, 7.5_0_6_4], [7.9_0_6_7, -1.3_3_6_9, 9.9_6_6_8], [9.0_7_1_2, -1.8_1_0_6, 7.0_3_8_0]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,A ,atol=1e-4 ) )
65
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Union[str, Any] ,A : Dict=7 ,A : Optional[int]=3 ,A : List[str]=18 ,A : Union[str, Any]=30 ,A : Tuple=400 ,A : Dict=True ,A : List[str]=None ,A : str=True ,A : Optional[Any]=False ,A : Optional[Any]=True ,A : List[str]=True ,A : Optional[int]=[0.5, 0.5, 0.5] ,A : List[str]=[0.5, 0.5, 0.5] ,): '''simple docstring''' UpperCAmelCase__ : str = parent UpperCAmelCase__ : List[str] = batch_size UpperCAmelCase__ : List[str] = num_channels UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : Optional[int] = max_resolution UpperCAmelCase__ : str = do_resize UpperCAmelCase__ : Tuple = size if size is not None else {"""height""": 18, """width""": 20} UpperCAmelCase__ : List[str] = do_thumbnail UpperCAmelCase__ : Optional[int] = do_align_axis UpperCAmelCase__ : Union[str, Any] = do_pad UpperCAmelCase__ : Tuple = do_normalize UpperCAmelCase__ : Optional[Any] = image_mean UpperCAmelCase__ : List[Any] = image_std def __lowercase ( self : Optional[int] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = DonutImageProcessor if is_vision_available() else None def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = DonutImageProcessingTester(self ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_thumbnail""" ) ) self.assertTrue(hasattr(A ,"""do_align_long_axis""" ) ) self.assertTrue(hasattr(A ,"""do_pad""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 18, """width""": 20} ) UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=(42, 84) ) self.assertEqual(image_processor.size ,{"""height""": 84, """width""": 42} ) def __lowercase ( self : Dict ): '''simple docstring''' pass @is_flaky() def __lowercase ( self : int ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase__ : int = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Tuple = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : List[str] ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase__ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Optional[int] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : Any ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : List[Any] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,)
65
1
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCAmelCase = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. __UpperCAmelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) __UpperCAmelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __UpperCAmelCase = re.compile(r'\[(.+?)\]\((https://huggingface\.co/.+?)\)') __UpperCAmelCase = { 'DecisionTransformerConfig', 'EncoderDecoderConfig', 'MusicgenConfig', 'RagConfig', 'SpeechEncoderDecoderConfig', 'TimmBackboneConfig', 'VisionEncoderDecoderConfig', 'VisionTextDualEncoderConfig', 'LlamaConfig', } def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = None # source code of `config_class` UpperCAmelCase__ : List[str] = inspect.getsource(__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = _re_checkpoint.findall(__UpperCamelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("""/""" ): UpperCAmelCase__ : str = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link UpperCAmelCase__ : List[str] = F"https://huggingface.co/{ckpt_name}" if ckpt_link == ckpt_link_from_name: UpperCAmelCase__ : List[str] = ckpt_name break return checkpoint def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : List[str] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue UpperCAmelCase__ : int = get_checkpoint_from_config_class(__UpperCamelCase ) UpperCAmelCase__ : List[str] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__UpperCamelCase ) if len(__UpperCamelCase ) > 0: UpperCAmelCase__ : str = """\n""".join(sorted(__UpperCamelCase ) ) raise ValueError(F"The following configurations don't contain any valid checkpoint:\n{message}" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
65
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class __lowercase ( __lowerCamelCase ): snake_case_ = """open-llama""" def __init__( self : Dict ,A : str=100_000 ,A : str=4_096 ,A : Optional[Any]=11_008 ,A : Tuple=32 ,A : str=32 ,A : Optional[int]="silu" ,A : List[Any]=2_048 ,A : str=0.0_2 ,A : Optional[int]=1e-6 ,A : int=True ,A : Tuple=0 ,A : str=1 ,A : Any=2 ,A : Optional[Any]=False ,A : int=True ,A : Any=0.1 ,A : Optional[Any]=0.1 ,A : Optional[Any]=True ,A : Union[str, Any]=True ,A : Tuple=None ,**A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : str = vocab_size UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : Tuple = intermediate_size UpperCAmelCase__ : Optional[int] = num_hidden_layers UpperCAmelCase__ : Any = num_attention_heads UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = initializer_range UpperCAmelCase__ : Optional[int] = rms_norm_eps UpperCAmelCase__ : Any = use_cache UpperCAmelCase__ : Optional[Any] = kwargs.pop( """use_memorry_efficient_attention""" ,A ) UpperCAmelCase__ : Any = hidden_dropout_prob UpperCAmelCase__ : str = attention_dropout_prob UpperCAmelCase__ : Optional[int] = use_stable_embedding UpperCAmelCase__ : Tuple = shared_input_output_embedding UpperCAmelCase__ : Tuple = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,tie_word_embeddings=A ,**A ,) def __lowercase ( self : Optional[Any] ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ f"got {self.rope_scaling}" ) UpperCAmelCase__ : List[Any] = self.rope_scaling.get("""type""" ,A ) UpperCAmelCase__ : int = self.rope_scaling.get("""factor""" ,A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(A ,A ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
65
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = {'configuration_ibert': ['IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'IBertConfig', 'IBertOnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'IBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'IBertForMaskedLM', 'IBertForMultipleChoice', 'IBertForQuestionAnswering', 'IBertForSequenceClassification', 'IBertForTokenClassification', 'IBertModel', 'IBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" from __future__ import annotations import math class __lowercase : def __init__( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : str = size # approximate the overall size of segment tree with given value UpperCAmelCase__ : Any = [0 for i in range(0 ,4 * size )] # create array to store lazy update UpperCAmelCase__ : Dict = [0 for i in range(0 ,4 * size )] UpperCAmelCase__ : List[Any] = [0 for i in range(0 ,4 * size )] # flag for lazy update def __lowercase ( self : int ,A : int ): '''simple docstring''' return idx * 2 def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' return idx * 2 + 1 def __lowercase ( self : Union[str, Any] ,A : int ,A : int ,A : int ,A : list[int] ): '''simple docstring''' if left_element == right_element: UpperCAmelCase__ : Optional[int] = a[left_element - 1] else: UpperCAmelCase__ : Dict = (left_element + right_element) // 2 self.build(self.left(A ) ,A ,A ,A ) self.build(self.right(A ) ,mid + 1 ,A ,A ) UpperCAmelCase__ : Union[str, Any] = max( self.segment_tree[self.left(A )] ,self.segment_tree[self.right(A )] ) def __lowercase ( self : Optional[int] ,A : int ,A : int ,A : int ,A : int ,A : int ,A : int ): '''simple docstring''' if self.flag[idx] is True: UpperCAmelCase__ : Dict = self.lazy[idx] UpperCAmelCase__ : int = False if left_element != right_element: UpperCAmelCase__ : List[str] = self.lazy[idx] UpperCAmelCase__ : List[Any] = self.lazy[idx] UpperCAmelCase__ : List[Any] = True UpperCAmelCase__ : Any = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: UpperCAmelCase__ : Any = val if left_element != right_element: UpperCAmelCase__ : str = val UpperCAmelCase__ : List[str] = val UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[int] = True return True UpperCAmelCase__ : Optional[int] = (left_element + right_element) // 2 self.update(self.left(A ) ,A ,A ,A ,A ,A ) self.update(self.right(A ) ,mid + 1 ,A ,A ,A ,A ) UpperCAmelCase__ : List[str] = max( self.segment_tree[self.left(A )] ,self.segment_tree[self.right(A )] ) return True def __lowercase ( self : List[str] ,A : int ,A : int ,A : int ,A : int ,A : int ): '''simple docstring''' if self.flag[idx] is True: UpperCAmelCase__ : Any = self.lazy[idx] UpperCAmelCase__ : Optional[int] = False if left_element != right_element: UpperCAmelCase__ : List[Any] = self.lazy[idx] UpperCAmelCase__ : Union[str, Any] = self.lazy[idx] UpperCAmelCase__ : Optional[int] = True UpperCAmelCase__ : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] UpperCAmelCase__ : Dict = (left_element + right_element) // 2 UpperCAmelCase__ : Tuple = self.query(self.left(A ) ,A ,A ,A ,A ) UpperCAmelCase__ : Any = self.query(self.right(A ) ,mid + 1 ,A ,A ,A ) return max(A ,A ) def __str__( self : str ): '''simple docstring''' return str([self.query(1 ,1 ,self.size ,A ,A ) for i in range(1 ,self.size + 1 )] ) if __name__ == "__main__": __UpperCAmelCase = [1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] __UpperCAmelCase = 15 __UpperCAmelCase = SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
65
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """attention_mask"""] def __init__( self : Any ,A : str=80 ,A : Optional[int]=16_000 ,A : int=0.0 ,A : str=10 ,A : Any=25 ,A : str="hamming_window" ,A : int=3_2_7_6_8.0 ,A : List[str]=0.9_7 ,A : Optional[int]=1.0 ,A : Optional[Any]=True ,A : Tuple=True ,A : Any=False ,**A : int ,): '''simple docstring''' super().__init__(feature_size=A ,sampling_rate=A ,padding_value=A ,**A ) UpperCAmelCase__ : str = feature_size UpperCAmelCase__ : int = sampling_rate UpperCAmelCase__ : int = padding_value UpperCAmelCase__ : Dict = hop_length UpperCAmelCase__ : int = win_length UpperCAmelCase__ : Dict = frame_signal_scale UpperCAmelCase__ : Dict = preemphasis_coeff UpperCAmelCase__ : str = mel_floor UpperCAmelCase__ : Any = normalize_means UpperCAmelCase__ : str = normalize_vars UpperCAmelCase__ : int = win_function UpperCAmelCase__ : List[Any] = return_attention_mask UpperCAmelCase__ : str = win_length * sampling_rate // 1_000 UpperCAmelCase__ : List[Any] = hop_length * sampling_rate // 1_000 UpperCAmelCase__ : int = optimal_fft_length(self.sample_size ) UpperCAmelCase__ : List[Any] = (self.n_fft // 2) + 1 def __lowercase ( self : Union[str, Any] ,A : np.array ): '''simple docstring''' if self.win_function == "hamming_window": UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=A ) else: UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ) UpperCAmelCase__ : Union[str, Any] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) UpperCAmelCase__ : Optional[Any] = spectrogram( one_waveform * self.frame_signal_scale ,window=A ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=A ,preemphasis=self.preemphasis_coeff ,mel_filters=A ,mel_floor=self.mel_floor ,log_mel="""log""" ,) return msfc_features.T def __lowercase ( self : str ,A : Any ,A : Optional[int] ,A : str ): '''simple docstring''' # make sure we normalize float32 arrays if self.normalize_means: UpperCAmelCase__ : Optional[Any] = x[:input_length].mean(axis=0 ) UpperCAmelCase__ : Any = np.subtract(A ,A ) if self.normalize_vars: UpperCAmelCase__ : str = x[:input_length].std(axis=0 ) UpperCAmelCase__ : Optional[int] = np.divide(A ,A ) if input_length < x.shape[0]: UpperCAmelCase__ : int = padding_value # make sure array is in float32 UpperCAmelCase__ : str = x.astype(np.floataa ) return x def __lowercase ( self : Union[str, Any] ,A : List[np.ndarray] ,A : Optional[np.ndarray] = None ): '''simple docstring''' UpperCAmelCase__ : Any = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(A ,A ,self.padding_value ) for x, n in zip(A ,A )] def __call__( self : Union[str, Any] ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : Union[bool, str, PaddingStrategy] = False ,A : Optional[int] = None ,A : bool = False ,A : Optional[int] = None ,A : Optional[bool] = None ,A : Optional[Union[str, TensorType]] = None ,A : Optional[int] = None ,**A : Tuple ,): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" f" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" f" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[Any] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : Any = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : List[str] = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Union[str, Any] = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [raw_speech] # extract fbank features UpperCAmelCase__ : Tuple = [self._extract_mfsc_features(A ) for one_waveform in raw_speech] # convert into correct format for padding UpperCAmelCase__ : str = BatchFeature({"""input_features""": features} ) UpperCAmelCase__ : Optional[Any] = self.pad( A ,padding=A ,max_length=A ,truncation=A ,pad_to_multiple_of=A ,return_attention_mask=A ,**A ,) # make sure list is in array format UpperCAmelCase__ : Tuple = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] ,A ): UpperCAmelCase__ : Union[str, Any] = [np.asarray(A ,dtype=np.floataa ) for feature in input_features] UpperCAmelCase__ : Dict = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: UpperCAmelCase__ : Union[str, Any] = ( np.array(A ,dtype=np.intaa ) if self._get_padding_strategies(A ,max_length=A ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) UpperCAmelCase__ : Any = self.normalize( padded_inputs["""input_features"""] ,attention_mask=A ) if return_tensors is not None: UpperCAmelCase__ : Union[str, Any] = padded_inputs.convert_to_tensors(A ) return padded_inputs
65
1
"""simple docstring""" import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = int(__UpperCamelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = t // 3600, (t // 60) % 60, t % 60 return F"{h}:{m:02d}:{s:02d}" if h != 0 else F"{m:02d}:{s:02d}" def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=300 ): '''simple docstring''' return F"\n <div>\n {prefix}\n <progress value='{value}' max='{total}' style='width:{width}px; height:20px; vertical-align: middle;'></progress>\n {label}\n </div>\n " def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = """<table border=\"1\" class=\"dataframe\">\n""" html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += F" <th>{i}</th>\n" html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: UpperCAmelCase__ : Optional[int] = F"{elt:.6f}" if isinstance(__UpperCamelCase , __UpperCamelCase ) else str(__UpperCamelCase ) html_code += F" <td>{elt}</td>\n" html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class __lowercase : snake_case_ = 5 snake_case_ = 0.2 def __init__( self : int ,A : int ,A : Optional[str] = None ,A : bool = True ,A : Optional["NotebookTrainingTracker"] = None ,A : int = 300 ,): '''simple docstring''' UpperCAmelCase__ : int = total UpperCAmelCase__ : Dict = """""" if prefix is None else prefix UpperCAmelCase__ : List[Any] = leave UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : str = width UpperCAmelCase__ : Optional[Any] = None UpperCAmelCase__ : int = None UpperCAmelCase__ : Dict = None def __lowercase ( self : Tuple ,A : int ,A : bool = False ,A : str = None ): '''simple docstring''' UpperCAmelCase__ : Dict = value if comment is not None: UpperCAmelCase__ : Optional[int] = comment if self.last_value is None: UpperCAmelCase__ : Optional[Any] = time.time() UpperCAmelCase__ : int = value UpperCAmelCase__ : Dict = None UpperCAmelCase__ : Any = self.warmup UpperCAmelCase__ : Tuple = 1 self.update_bar(A ) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for ,self.total ): if self.first_calls > 0: self.first_calls -= 1 UpperCAmelCase__ : int = time.time() UpperCAmelCase__ : str = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: UpperCAmelCase__ : Union[str, Any] = self.elapsed_time / (value - self.start_value) else: UpperCAmelCase__ : List[Any] = None if value >= self.total: UpperCAmelCase__ : Optional[int] = self.total UpperCAmelCase__ : List[Any] = None if not self.leave: self.close() elif self.average_time_per_item is not None: UpperCAmelCase__ : Optional[Any] = self.average_time_per_item * (self.total - value) self.update_bar(A ) UpperCAmelCase__ : Tuple = value UpperCAmelCase__ : List[str] = current_time if self.average_time_per_item is None: UpperCAmelCase__ : List[str] = 1 else: UpperCAmelCase__ : Any = max(int(self.update_every / self.average_time_per_item ) ,1 ) def __lowercase ( self : str ,A : int ,A : List[str]=None ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = """ """ * (len(str(self.total ) ) - len(str(A ) )) + str(A ) if self.elapsed_time is None: UpperCAmelCase__ : Tuple = f"[{spaced_value}/{self.total} : < :" elif self.predicted_remaining is None: UpperCAmelCase__ : Any = f"[{spaced_value}/{self.total} {format_time(self.elapsed_time )}" else: UpperCAmelCase__ : List[Any] = ( f"[{spaced_value}/{self.total} {format_time(self.elapsed_time )} <" f" {format_time(self.predicted_remaining )}" ) self.label += f", {1/self.average_time_per_item:.2f} it/s" self.label += "]" if self.comment is None or len(self.comment ) == 0 else f", {self.comment}]" self.display() def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = html_progress_bar(self.value ,self.total ,self.prefix ,self.label ,self.width ) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: UpperCAmelCase__ : str = disp.display(disp.HTML(self.html_code ) ,display_id=A ) else: self.output.update(disp.HTML(self.html_code ) ) def __lowercase ( self : Optional[int] ): '''simple docstring''' if self.parent is None and self.output is not None: self.output.update(disp.HTML("""""" ) ) class __lowercase ( __lowerCamelCase ): def __init__( self : Dict ,A : Optional[int] ,A : Dict=None ): '''simple docstring''' super().__init__(A ) UpperCAmelCase__ : Union[str, Any] = None if column_names is None else [column_names] UpperCAmelCase__ : Union[str, Any] = None def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Tuple = html_progress_bar(self.value ,self.total ,self.prefix ,self.label ,self.width ) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table ) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: UpperCAmelCase__ : Dict = disp.display(disp.HTML(self.html_code ) ,display_id=A ) else: self.output.update(disp.HTML(self.html_code ) ) def __lowercase ( self : Any ,A : List[Any] ): '''simple docstring''' if self.inner_table is None: UpperCAmelCase__ : Union[str, Any] = [list(values.keys() ), list(values.values() )] else: UpperCAmelCase__ : List[str] = self.inner_table[0] if len(self.inner_table ) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(A ) UpperCAmelCase__ : Dict = columns self.inner_table.append([values[c] for c in columns] ) def __lowercase ( self : List[Any] ,A : int ,A : Dict=None ,A : str=300 ): '''simple docstring''' UpperCAmelCase__ : List[str] = NotebookProgressBar(A ,prefix=A ,parent=self ,width=A ) return self.child_bar def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = None self.display() class __lowercase ( __lowerCamelCase ): def __init__( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = None UpperCAmelCase__ : Any = None UpperCAmelCase__ : Any = False def __lowercase ( self : Tuple ,A : Optional[Any] ,A : int ,A : str ,**A : int ): '''simple docstring''' UpperCAmelCase__ : Any = """Epoch""" if args.evaluation_strategy == IntervalStrategy.EPOCH else """Step""" UpperCAmelCase__ : Any = 0 UpperCAmelCase__ : List[str] = 0 UpperCAmelCase__ : int = [self.first_column] + ["""Training Loss"""] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append("""Validation Loss""" ) UpperCAmelCase__ : Tuple = NotebookTrainingTracker(state.max_steps ,A ) def __lowercase ( self : List[Any] ,A : Optional[Any] ,A : List[str] ,A : int ,**A : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = int(state.epoch ) if int(state.epoch ) == state.epoch else f"{state.epoch:.2f}" self.training_tracker.update( state.global_step + 1 ,comment=f"Epoch {epoch}/{state.num_train_epochs}" ,force_update=self._force_next_update ,) UpperCAmelCase__ : Any = False def __lowercase ( self : Tuple ,A : Dict ,A : Union[str, Any] ,A : List[Any] ,A : List[str]=None ,**A : Optional[Any] ): '''simple docstring''' if not has_length(A ): return if self.prediction_bar is None: if self.training_tracker is not None: UpperCAmelCase__ : Optional[int] = self.training_tracker.add_child(len(A ) ) else: UpperCAmelCase__ : int = NotebookProgressBar(len(A ) ) self.prediction_bar.update(1 ) else: self.prediction_bar.update(self.prediction_bar.value + 1 ) def __lowercase ( self : List[str] ,A : Optional[int] ,A : Any ,A : Union[str, Any] ,**A : Optional[int] ): '''simple docstring''' if self.prediction_bar is not None: self.prediction_bar.close() UpperCAmelCase__ : List[str] = None def __lowercase ( self : List[str] ,A : Any ,A : Optional[Any] ,A : List[Any] ,A : Optional[Any]=None ,**A : str ): '''simple docstring''' # Only for when there is no evaluation if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: UpperCAmelCase__ : List[Any] = {"""Training Loss""": logs["""loss"""]} # First column is necessarily Step sine we're not in epoch eval strategy UpperCAmelCase__ : str = state.global_step self.training_tracker.write_line(A ) def __lowercase ( self : Optional[Any] ,A : Tuple ,A : List[str] ,A : int ,A : str=None ,**A : List[Any] ): '''simple docstring''' if self.training_tracker is not None: UpperCAmelCase__ : List[Any] = {"""Training Loss""": """No log""", """Validation Loss""": """No log"""} for log in reversed(state.log_history ): if "loss" in log: UpperCAmelCase__ : Optional[int] = log["""loss"""] break if self.first_column == "Epoch": UpperCAmelCase__ : Dict = int(state.epoch ) else: UpperCAmelCase__ : List[Any] = state.global_step UpperCAmelCase__ : List[Any] = """eval""" for k in metrics: if k.endswith("""_loss""" ): UpperCAmelCase__ : Optional[int] = re.sub(R"""\_loss$""" ,"""""" ,A ) UpperCAmelCase__ : Dict = metrics.pop("""total_flos""" ,A ) UpperCAmelCase__ : str = metrics.pop("""epoch""" ,A ) UpperCAmelCase__ : Union[str, Any] = metrics.pop(f"{metric_key_prefix}_runtime" ,A ) UpperCAmelCase__ : Any = metrics.pop(f"{metric_key_prefix}_samples_per_second" ,A ) UpperCAmelCase__ : int = metrics.pop(f"{metric_key_prefix}_steps_per_second" ,A ) UpperCAmelCase__ : str = metrics.pop(f"{metric_key_prefix}_jit_compilation_time" ,A ) for k, v in metrics.items(): if k == f"{metric_key_prefix}_loss": UpperCAmelCase__ : Tuple = v else: UpperCAmelCase__ : int = k.split("""_""" ) UpperCAmelCase__ : Tuple = """ """.join([part.capitalize() for part in splits[1:]] ) UpperCAmelCase__ : Optional[int] = v self.training_tracker.write_line(A ) self.training_tracker.remove_child() UpperCAmelCase__ : List[Any] = None # Evaluation takes a long time so we should force the next update. UpperCAmelCase__ : List[str] = True def __lowercase ( self : Dict ,A : Optional[Any] ,A : Optional[int] ,A : int ,**A : int ): '''simple docstring''' self.training_tracker.update( state.global_step ,comment=f"Epoch {int(state.epoch )}/{state.num_train_epochs}" ,force_update=A ) UpperCAmelCase__ : List[Any] = None
65
"""simple docstring""" from math import factorial def lowerCAmelCase ( __UpperCamelCase = 100 ): '''simple docstring''' return sum(int(__UpperCamelCase ) for x in str(factorial(__UpperCamelCase ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
65
1
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
65
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Optional[int] ,A : int=13 ,A : Tuple=7 ,A : Dict=True ,A : Optional[int]=True ,A : Tuple=True ,A : str=True ,A : Any=99 ,A : Tuple=32 ,A : Dict=5 ,A : Optional[int]=4 ,A : Dict=37 ,A : Any="gelu" ,A : Any=0.1 ,A : Optional[int]=0.1 ,A : Union[str, Any]=512 ,A : Any=16 ,A : List[str]=2 ,A : List[Any]=0.0_2 ,A : Optional[int]=4 ,): '''simple docstring''' UpperCAmelCase__ : Dict = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : Optional[int] = is_training UpperCAmelCase__ : Optional[Any] = use_attention_mask UpperCAmelCase__ : int = use_token_type_ids UpperCAmelCase__ : int = use_labels UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : List[Any] = type_vocab_size UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : List[Any] = num_choices def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : List[str] = None if self.use_attention_mask: UpperCAmelCase__ : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : int = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=A ,) return config, input_ids, attention_mask def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = config_and_inputs UpperCAmelCase__ : str = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = FlaxDistilBertModelTester(self ) @slow def __lowercase ( self : Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = model_class_name.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(A ) @require_flax class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) UpperCAmelCase__ : str = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ : Dict = model(A ,attention_mask=A )[0] UpperCAmelCase__ : List[Any] = (1, 11, 768) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Any = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,A ,atol=1e-4 ) )
65
1
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase=0 ): '''simple docstring''' return sorted(__UpperCamelCase , key=lambda __UpperCamelCase : x[column] ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=float("""inf""" ) ): '''simple docstring''' for i in range(points_counts - 1 ): for j in range(i + 1 , __UpperCamelCase ): UpperCAmelCase__ : Optional[int] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: UpperCAmelCase__ : Any = current_dis return min_dis def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=float("""inf""" ) ): '''simple docstring''' for i in range(min(6 , points_counts - 1 ) , __UpperCamelCase ): for j in range(max(0 , i - 6 ) , __UpperCamelCase ): UpperCAmelCase__ : List[Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: UpperCAmelCase__ : List[Any] = current_dis return min_dis def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if points_counts <= 3: return dis_between_closest_pair(__UpperCamelCase , __UpperCamelCase ) # recursion UpperCAmelCase__ : Any = points_counts // 2 UpperCAmelCase__ : Union[str, Any] = closest_pair_of_points_sqr( __UpperCamelCase , points_sorted_on_y[:mid] , __UpperCamelCase ) UpperCAmelCase__ : Optional[int] = closest_pair_of_points_sqr( __UpperCamelCase , points_sorted_on_y[mid:] , points_counts - mid ) UpperCAmelCase__ : Dict = min(__UpperCamelCase , __UpperCamelCase ) UpperCAmelCase__ : List[Any] = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(__UpperCamelCase ) UpperCAmelCase__ : Tuple = dis_between_closest_in_strip( __UpperCamelCase , len(__UpperCamelCase ) , __UpperCamelCase ) return min(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = column_based_sort(__UpperCamelCase , column=0 ) UpperCAmelCase__ : Optional[int] = column_based_sort(__UpperCamelCase , column=1 ) return ( closest_pair_of_points_sqr( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) ) ** 0.5 if __name__ == "__main__": __UpperCAmelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print('Distance:', closest_pair_of_points(points, len(points)))
65
"""simple docstring""" __UpperCAmelCase = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image', 'mask_image']) __UpperCAmelCase = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['example_image', 'image', 'mask_image']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset(['input_tokens']) __UpperCAmelCase = frozenset(['input_tokens'])
65
1
"""simple docstring""" import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin __UpperCAmelCase = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class __lowercase : def __init__( self : Optional[int] ,A : Optional[int] ,A : Optional[int]=16 ,A : Optional[Any]=13 ,A : Optional[Any]=7 ,A : List[str]=14 ,A : List[str]=10 ,A : Tuple=19 ,A : Optional[int]=5 ,A : str=4 ,A : Union[str, Any]=True ,A : int=16 ,A : int=2 ,A : List[Any]=4 ,A : Tuple=4 ,A : int="gelu" ,A : Union[str, Any]=0.1 ,A : List[str]=0.1 ,A : Dict=[1, 2, 3, 4, 5] ,A : int=25 ,A : Dict=5 ,): '''simple docstring''' UpperCAmelCase__ : List[Any] = d_model UpperCAmelCase__ : str = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : Any = prediction_length UpperCAmelCase__ : List[str] = context_length UpperCAmelCase__ : Optional[int] = cardinality UpperCAmelCase__ : Dict = num_time_features UpperCAmelCase__ : Tuple = lags_sequence UpperCAmelCase__ : str = embedding_dimension UpperCAmelCase__ : int = is_training UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : Union[str, Any] = num_hidden_layers UpperCAmelCase__ : Optional[int] = num_attention_heads UpperCAmelCase__ : List[Any] = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : Tuple = hidden_dropout_prob UpperCAmelCase__ : Tuple = attention_probs_dropout_prob UpperCAmelCase__ : Any = context_length UpperCAmelCase__ : Optional[Any] = prediction_length + label_length UpperCAmelCase__ : List[Any] = label_length UpperCAmelCase__ : str = moving_average UpperCAmelCase__ : Dict = autocorrelation_factor def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return AutoformerConfig( d_model=self.d_model ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,prediction_length=self.prediction_length ,context_length=self.context_length ,label_length=self.label_length ,lags_sequence=self.lags_sequence ,num_time_features=self.num_time_features ,num_static_categorical_features=1 ,cardinality=[self.cardinality] ,embedding_dimension=[self.embedding_dimension] ,moving_average=self.moving_average ,) def __lowercase ( self : Tuple ,A : Any ): '''simple docstring''' UpperCAmelCase__ : List[Any] = config.context_length + max(config.lags_sequence ) UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, 1] ,config.cardinality[0] ) UpperCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, _past_length] ) UpperCAmelCase__ : Any = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs UpperCAmelCase__ : str = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) UpperCAmelCase__ : int = floats_tensor([self.batch_size, config.prediction_length] ) UpperCAmelCase__ : List[Any] = { """past_values""": past_values, """static_categorical_features""": static_categorical_features, """past_time_features""": past_time_features, """past_observed_mask""": past_observed_mask, """future_time_features""": future_time_features, """future_values""": future_values, } return inputs_dict def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_config() UpperCAmelCase__ : List[str] = self.prepare_autoformer_inputs_dict(A ) return config, inputs_dict def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = self.prepare_config_and_inputs() return config, inputs_dict def __lowercase ( self : int ,A : Optional[int] ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : str = AutoformerModel(config=A ).to(A ).eval() UpperCAmelCase__ : List[Any] = model(**A ) UpperCAmelCase__ : List[str] = outputs.encoder_last_hidden_state UpperCAmelCase__ : int = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ : Any = model.get_encoder() encoder.save_pretrained(A ) UpperCAmelCase__ : Optional[int] = AutoformerEncoder.from_pretrained(A ).to(A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = model.create_network_inputs(**A ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) UpperCAmelCase__ : List[str] = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) ,dim=-1 ,) UpperCAmelCase__ : Union[str, Any] = encoder(inputs_embeds=A )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) UpperCAmelCase__ : Dict = ( torch.mean(transformer_inputs[:, : config.context_length, ...] ,dim=1 ) .unsqueeze(1 ) .repeat(1 ,config.prediction_length ,1 ) ) UpperCAmelCase__ : Dict = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] ,device=enc_input.device ,) UpperCAmelCase__ : str = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) ,dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) ,dim=-1 ,) UpperCAmelCase__ : List[str] = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) ,dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) ,dim=-1 ,) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ : List[Any] = model.get_decoder() decoder.save_pretrained(A ) UpperCAmelCase__ : Dict = AutoformerDecoder.from_pretrained(A ).to(A ) UpperCAmelCase__ : Any = decoder( trend=A ,inputs_embeds=A ,encoder_hidden_states=A ,)[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class __lowercase ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): snake_case_ = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () snake_case_ = (AutoformerForPrediction,) if is_torch_available() else () snake_case_ = {"""feature-extraction""": AutoformerModel} if is_torch_available() else {} snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Tuple = AutoformerModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self ,config_class=A ,has_text_modality=A ) def __lowercase ( self : str ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(A ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A ) UpperCAmelCase__ , UpperCAmelCase__ : Dict = model_class.from_pretrained(A ,output_loading_info=A ) self.assertEqual(info["""missing_keys"""] ,[] ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*A ) @unittest.skip(reason="""Model has no tokens embeddings""" ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' pass def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : List[str] = inspect.signature(getattr(A ,"""forward""" ) ) # The main input is the name of the argument after `self` UpperCAmelCase__ : Optional[int] = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name ,A ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[Any] = model_class(A ) UpperCAmelCase__ : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : List[Any] = [*signature.parameters.keys()] UpperCAmelCase__ : List[Any] = [ """past_values""", """past_time_features""", """past_observed_mask""", """static_categorical_features""", """static_real_features""", """future_values""", """future_time_features""", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("""future_observed_mask""" ) expected_arg_names.extend( [ """decoder_attention_mask""", """head_mask""", """decoder_head_mask""", """cross_attn_head_mask""", """encoder_outputs""", """past_key_values""", """output_hidden_states""", """output_attentions""", """use_cache""", """return_dict""", ] ) self.assertListEqual(arg_names[: len(A )] ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : str = True UpperCAmelCase__ : str = getattr(self.model_tester ,"""seq_length""" ,A ) UpperCAmelCase__ : int = getattr(self.model_tester ,"""decoder_seq_length""" ,A ) UpperCAmelCase__ : str = getattr(self.model_tester ,"""encoder_seq_length""" ,A ) UpperCAmelCase__ : int = getattr(self.model_tester ,"""d_model""" ,A ) UpperCAmelCase__ : Union[str, Any] = getattr(self.model_tester ,"""num_attention_heads""" ,A ) UpperCAmelCase__ : List[Any] = d_model // num_attention_heads for model_class in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : str = True UpperCAmelCase__ : Optional[int] = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(**self._prepare_for_class(A ,A ) ) UpperCAmelCase__ : Any = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(A ) ,self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase__ : int = True UpperCAmelCase__ : List[Any] = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(**self._prepare_for_class(A ,A ) ) UpperCAmelCase__ : List[str] = outputs.encoder_attentions self.assertEqual(len(A ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, encoder_seq_length, dim] ,) UpperCAmelCase__ : List[str] = len(A ) UpperCAmelCase__ : Dict = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(A ,A ) # decoder attentions UpperCAmelCase__ : int = outputs.decoder_attentions self.assertIsInstance(A ,(list, tuple) ) self.assertEqual(len(A ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, decoder_seq_length, dim] ,) # cross attentions UpperCAmelCase__ : int = outputs.cross_attentions self.assertIsInstance(A ,(list, tuple) ) self.assertEqual(len(A ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, decoder_seq_length, dim] ,) # Check attention is always last and order is fine UpperCAmelCase__ : List[Any] = True UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : Any = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): UpperCAmelCase__ : List[str] = model(**self._prepare_for_class(A ,A ) ) self.assertEqual(out_len + 2 ,len(A ) ) UpperCAmelCase__ : Tuple = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(A ) ,self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_attention_heads, encoder_seq_length, dim] ,) @is_flaky() def __lowercase ( self : Tuple ): '''simple docstring''' super().test_retain_grad_hidden_states_attentions() def lowerCAmelCase ( __UpperCamelCase="train-batch.pt" ): '''simple docstring''' UpperCAmelCase__ : Any = hf_hub_download(repo_id="""hf-internal-testing/tourism-monthly-batch""" , filename=__UpperCamelCase , repo_type="""dataset""" ) UpperCAmelCase__ : Dict = torch.load(__UpperCamelCase , map_location=__UpperCamelCase ) return batch @require_torch @slow class __lowercase ( unittest.TestCase ): def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Any = AutoformerModel.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(A ) UpperCAmelCase__ : Dict = prepare_batch() with torch.no_grad(): UpperCAmelCase__ : Any = model( past_values=batch["""past_values"""] ,past_time_features=batch["""past_time_features"""] ,past_observed_mask=batch["""past_observed_mask"""] ,static_categorical_features=batch["""static_categorical_features"""] ,future_values=batch["""future_values"""] ,future_time_features=batch["""future_time_features"""] ,)[0] UpperCAmelCase__ : Dict = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : str = torch.tensor( [[0.3_5_9_3, -1.3_3_9_8, 0.6_3_3_0], [0.2_2_7_9, 1.5_3_9_6, -0.1_7_9_2], [0.0_4_5_0, 1.3_2_2_5, -0.2_3_3_5]] ,device=A ) self.assertTrue(torch.allclose(output[0, :3, :3] ,A ,atol=A ) ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(A ) UpperCAmelCase__ : Tuple = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): UpperCAmelCase__ : List[Any] = model( past_values=batch["""past_values"""] ,past_time_features=batch["""past_time_features"""] ,past_observed_mask=batch["""past_observed_mask"""] ,static_categorical_features=batch["""static_categorical_features"""] ,).encoder_last_hidden_state UpperCAmelCase__ : Tuple = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Optional[int] = torch.tensor( [[-0.0_7_3_4, -0.9_0_3_6, 0.8_3_5_8], [4.7_1_8_6, 2.4_1_1_3, 1.9_5_8_1], [1.7_9_5_3, 2.3_5_5_8, 1.2_9_7_0]] ,device=A ) self.assertTrue(torch.allclose(output[0, :3, :3] ,A ,atol=A ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(A ) UpperCAmelCase__ : Any = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): UpperCAmelCase__ : List[Any] = model.generate( static_categorical_features=batch["""static_categorical_features"""] ,past_time_features=batch["""past_time_features"""] ,past_values=batch["""past_values"""] ,future_time_features=batch["""future_time_features"""] ,past_observed_mask=batch["""past_observed_mask"""] ,) UpperCAmelCase__ : List[Any] = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape ,A ) UpperCAmelCase__ : Optional[int] = torch.tensor([3_1_3_0.6_7_6_3, 4_0_5_6.5_2_9_3, 7_0_5_3.0_7_8_6] ,device=A ) UpperCAmelCase__ : Any = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] ,A ,rtol=1e-1 ) )
65
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase__ : Tuple = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[Any] = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase__ : int = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 16_000, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase__ : Optional[int] = tempfile.mkdtemp() UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,A ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.feature_extraction_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) # load decoder from hub UpperCAmelCase__ : int = """hf-internal-testing/ngram-beam-search-decoder""" def __lowercase ( self : str ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.add_kwargs_tokens_map.copy() kwargs.update(A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[str] ,**A : Dict ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Any ,**A : List[Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name ,**A ) def __lowercase ( self : Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_decoder() UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() ,feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor ,A ) # decoder self.assertEqual(processor.decoder._alphabet.labels ,decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set ,decoder.model_container[decoder._model_key]._unigram_set ,) self.assertIsInstance(processor.decoder ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname ,alpha=5.0 ,beta=3.0 ,score_boundary=-7.0 ,unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha ,5.0 ) self.assertEqual(processor.language_model.beta ,3.0 ) self.assertEqual(processor.language_model.score_boundary ,-7.0 ) self.assertEqual(processor.language_model.unk_score_offset ,3 ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(A ,"""include""" ): WavaVecaProcessorWithLM( tokenizer=A ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : Optional[Any] = feature_extractor(A ,return_tensors="""np""" ) UpperCAmelCase__ : List[Any] = processor(A ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : List[Any] = """This is a test string""" UpperCAmelCase__ : int = processor(text=A ) UpperCAmelCase__ : Dict = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __lowercase ( self : Tuple ,A : List[Any]=(2, 10, 16) ,A : Dict=77 ): '''simple docstring''' np.random.seed(A ) return np.random.rand(*A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits(shape=(10, 16) ,seed=13 ) UpperCAmelCase__ : Tuple = processor.decode(A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(A )[0] self.assertEqual(decoded_decoder[0] ,decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" ,decoded_processor.text ) self.assertEqual(decoded_decoder[-2] ,decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] ,decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def __lowercase ( self : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_feature_extractor() UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : List[str] = processor.batch_decode(A ) else: with get_context(A ).Pool() as pool: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(A ,A ) UpperCAmelCase__ : Optional[Any] = list(A ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A ,decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] ,decoded_processor.text ) self.assertListEqual(A ,decoded_processor.logit_score ) self.assertListEqual(A ,decoded_processor.lm_score ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits() UpperCAmelCase__ : Any = 15 UpperCAmelCase__ : Dict = -2_0.0 UpperCAmelCase__ : List[Any] = -4.0 UpperCAmelCase__ : Union[str, Any] = processor.batch_decode( A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[str] = decoded_processor_out.text UpperCAmelCase__ : List[str] = list(A ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( A ,A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] ,A ) self.assertTrue(np.array_equal(A ,decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] ,A ,atol=1e-3 ) ) self.assertTrue(np.array_equal(A ,decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] ,A ,atol=1e-3 ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits() UpperCAmelCase__ : Tuple = 2.0 UpperCAmelCase__ : str = 5.0 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = processor.batch_decode( A ,alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) UpperCAmelCase__ : Any = decoded_processor_out.text UpperCAmelCase__ : Union[str, Any] = list(A ) decoder.reset_params( alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch( A ,A ,) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] ,A ) UpperCAmelCase__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha ,2.0 ) self.assertEqual(lm_model.beta ,5.0 ) self.assertEqual(lm_model.unk_score_offset ,-2_0.0 ) self.assertEqual(lm_model.score_boundary ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : str = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Any = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[int] = os.listdir(A ) UpperCAmelCase__ : List[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(A ) UpperCAmelCase__ : Tuple = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[int] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Tuple = os.listdir(A ) UpperCAmelCase__ : Dict = os.listdir(A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = floats_list((3, 1_000) ) UpperCAmelCase__ : List[str] = processor_wavaveca(A ,return_tensors="""np""" ) UpperCAmelCase__ : Dict = processor_auto(A ,return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() ,input_auto[key].sum() ,delta=1e-2 ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Tuple = processor_wavaveca.batch_decode(A ) UpperCAmelCase__ : List[str] = processor_auto.batch_decode(A ) self.assertListEqual(decoded_wavaveca.text ,decoded_auto.text ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : int = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) self.assertListEqual( processor.model_input_names ,feature_extractor.model_input_names ,msg="""`processor` and `feature_extractor` model input names do not match""" ,) @staticmethod def __lowercase ( A : Optional[Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = self._get_dummy_logits()[0] UpperCAmelCase__ : List[str] = processor.decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ) ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""end_offset""" ) ,[1, 3, 5] ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : int = self._get_dummy_logits() UpperCAmelCase__ : Any = processor.batch_decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(A ,"""word""" ) ) for o in outputs["""word_offsets"""]] ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""end_offset""" ) ,[1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self : Tuple ): '''simple docstring''' import torch UpperCAmelCase__ : Any = load_dataset("""common_voice""" ,"""en""" ,split="""train""" ,streaming=A ) UpperCAmelCase__ : Tuple = ds.cast_column("""audio""" ,datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Tuple = iter(A ) UpperCAmelCase__ : Optional[int] = next(A ) UpperCAmelCase__ : List[Any] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase__ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Tuple = processor(sample["""audio"""]["""array"""] ,return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(A ).logits.cpu().numpy() UpperCAmelCase__ : Any = processor.decode(logits[0] ,output_word_offsets=A ) UpperCAmelCase__ : str = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase__ : Dict = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,A ) self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,output.text ) # output times UpperCAmelCase__ : str = torch.tensor(self.get_from_offsets(A ,"""start_time""" ) ) UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(A ,"""end_time""" ) ) # fmt: off UpperCAmelCase__ : Union[str, Any] = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : List[Any] = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) ) self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) )
65
1
"""simple docstring""" import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class __lowercase ( __lowerCamelCase ): snake_case_ = (UnCLIPScheduler,) def __lowercase ( self : Optional[int] ,**A : int ): '''simple docstring''' UpperCAmelCase__ : int = { """num_train_timesteps""": 1_000, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**A ) return config def __lowercase ( self : List[Any] ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=A ) def __lowercase ( self : int ): '''simple docstring''' for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=A ) def __lowercase ( self : Tuple ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=A ) def __lowercase ( self : int ): '''simple docstring''' for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=A ) def __lowercase ( self : int ): '''simple docstring''' for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=A ,prev_timestep=A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase__ : Tuple = self.get_scheduler_config(variance_type="""fixed_small_log""" ) UpperCAmelCase__ : Any = scheduler_class(**A ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0_000e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_5_4_9_6_2_5 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.9_9_9_4_9_8_7 ) ) < 1e-5 def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : int = self.scheduler_classes[0] UpperCAmelCase__ : Dict = self.get_scheduler_config(variance_type="""learned_range""" ) UpperCAmelCase__ : Tuple = scheduler_class(**A ) UpperCAmelCase__ : List[Any] = 0.5 assert scheduler._get_variance(1 ,predicted_variance=A ) - -1_0.1_7_1_2_7_9_0 < 1e-5 assert scheduler._get_variance(487 ,predicted_variance=A ) - -5.7_9_9_8_0_5_2 < 1e-5 assert scheduler._get_variance(999 ,predicted_variance=A ) - -0.0_0_1_0_0_1_1 < 1e-5 def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Any = self.scheduler_classes[0] UpperCAmelCase__ : Any = self.get_scheduler_config() UpperCAmelCase__ : Optional[Any] = scheduler_class(**A ) UpperCAmelCase__ : List[Any] = scheduler.timesteps UpperCAmelCase__ : Dict = self.dummy_model() UpperCAmelCase__ : List[Any] = self.dummy_sample_deter UpperCAmelCase__ : Optional[int] = torch.manual_seed(0 ) for i, t in enumerate(A ): # 1. predict noise residual UpperCAmelCase__ : Tuple = model(A ,A ) # 2. predict previous mean of sample x_t-1 UpperCAmelCase__ : Optional[int] = scheduler.step(A ,A ,A ,generator=A ).prev_sample UpperCAmelCase__ : Union[str, Any] = pred_prev_sample UpperCAmelCase__ : int = torch.sum(torch.abs(A ) ) UpperCAmelCase__ : Optional[Any] = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 2_5_2.2_6_8_2_4_9_5 ) < 1e-2 assert abs(result_mean.item() - 0.3_2_8_4_7_4_3 ) < 1e-3 def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase__ : Union[str, Any] = self.get_scheduler_config() UpperCAmelCase__ : int = scheduler_class(**A ) scheduler.set_timesteps(25 ) UpperCAmelCase__ : Optional[Any] = scheduler.timesteps UpperCAmelCase__ : Optional[int] = self.dummy_model() UpperCAmelCase__ : Union[str, Any] = self.dummy_sample_deter UpperCAmelCase__ : Tuple = torch.manual_seed(0 ) for i, t in enumerate(A ): # 1. predict noise residual UpperCAmelCase__ : Dict = model(A ,A ) if i + 1 == timesteps.shape[0]: UpperCAmelCase__ : Tuple = None else: UpperCAmelCase__ : Union[str, Any] = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 UpperCAmelCase__ : int = scheduler.step( A ,A ,A ,prev_timestep=A ,generator=A ).prev_sample UpperCAmelCase__ : str = pred_prev_sample UpperCAmelCase__ : Dict = torch.sum(torch.abs(A ) ) UpperCAmelCase__ : str = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 2_5_8.2_0_4_4_9_8_3 ) < 1e-2 assert abs(result_mean.item() - 0.3_3_6_2_0_3_8 ) < 1e-3 def __lowercase ( self : int ): '''simple docstring''' pass def __lowercase ( self : Dict ): '''simple docstring''' pass
65
"""simple docstring""" from sklearn.metrics import fa_score import datasets __UpperCAmelCase = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __UpperCAmelCase = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __UpperCAmelCase = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def __lowercase ( self : List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) ,reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"""] ,) def __lowercase ( self : Union[str, Any] ,A : List[str] ,A : List[Any] ,A : Optional[Any]=None ,A : List[str]=1 ,A : Optional[Any]="binary" ,A : Any=None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = fa_score( A ,A ,labels=A ,pos_label=A ,average=A ,sample_weight=A ) return {"f1": float(A ) if score.size == 1 else score}
65
1
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase__ : Tuple = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[Any] = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase__ : int = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 16_000, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase__ : Optional[int] = tempfile.mkdtemp() UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,A ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.feature_extraction_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) # load decoder from hub UpperCAmelCase__ : int = """hf-internal-testing/ngram-beam-search-decoder""" def __lowercase ( self : str ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.add_kwargs_tokens_map.copy() kwargs.update(A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[str] ,**A : Dict ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Any ,**A : List[Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name ,**A ) def __lowercase ( self : Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_decoder() UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() ,feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor ,A ) # decoder self.assertEqual(processor.decoder._alphabet.labels ,decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set ,decoder.model_container[decoder._model_key]._unigram_set ,) self.assertIsInstance(processor.decoder ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname ,alpha=5.0 ,beta=3.0 ,score_boundary=-7.0 ,unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha ,5.0 ) self.assertEqual(processor.language_model.beta ,3.0 ) self.assertEqual(processor.language_model.score_boundary ,-7.0 ) self.assertEqual(processor.language_model.unk_score_offset ,3 ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(A ,"""include""" ): WavaVecaProcessorWithLM( tokenizer=A ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : Optional[Any] = feature_extractor(A ,return_tensors="""np""" ) UpperCAmelCase__ : List[Any] = processor(A ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : List[Any] = """This is a test string""" UpperCAmelCase__ : int = processor(text=A ) UpperCAmelCase__ : Dict = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __lowercase ( self : Tuple ,A : List[Any]=(2, 10, 16) ,A : Dict=77 ): '''simple docstring''' np.random.seed(A ) return np.random.rand(*A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits(shape=(10, 16) ,seed=13 ) UpperCAmelCase__ : Tuple = processor.decode(A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(A )[0] self.assertEqual(decoded_decoder[0] ,decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" ,decoded_processor.text ) self.assertEqual(decoded_decoder[-2] ,decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] ,decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def __lowercase ( self : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_feature_extractor() UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : List[str] = processor.batch_decode(A ) else: with get_context(A ).Pool() as pool: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(A ,A ) UpperCAmelCase__ : Optional[Any] = list(A ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A ,decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] ,decoded_processor.text ) self.assertListEqual(A ,decoded_processor.logit_score ) self.assertListEqual(A ,decoded_processor.lm_score ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits() UpperCAmelCase__ : Any = 15 UpperCAmelCase__ : Dict = -2_0.0 UpperCAmelCase__ : List[Any] = -4.0 UpperCAmelCase__ : Union[str, Any] = processor.batch_decode( A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[str] = decoded_processor_out.text UpperCAmelCase__ : List[str] = list(A ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( A ,A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] ,A ) self.assertTrue(np.array_equal(A ,decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] ,A ,atol=1e-3 ) ) self.assertTrue(np.array_equal(A ,decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] ,A ,atol=1e-3 ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits() UpperCAmelCase__ : Tuple = 2.0 UpperCAmelCase__ : str = 5.0 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = processor.batch_decode( A ,alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) UpperCAmelCase__ : Any = decoded_processor_out.text UpperCAmelCase__ : Union[str, Any] = list(A ) decoder.reset_params( alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch( A ,A ,) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] ,A ) UpperCAmelCase__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha ,2.0 ) self.assertEqual(lm_model.beta ,5.0 ) self.assertEqual(lm_model.unk_score_offset ,-2_0.0 ) self.assertEqual(lm_model.score_boundary ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : str = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Any = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[int] = os.listdir(A ) UpperCAmelCase__ : List[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(A ) UpperCAmelCase__ : Tuple = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[int] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Tuple = os.listdir(A ) UpperCAmelCase__ : Dict = os.listdir(A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = floats_list((3, 1_000) ) UpperCAmelCase__ : List[str] = processor_wavaveca(A ,return_tensors="""np""" ) UpperCAmelCase__ : Dict = processor_auto(A ,return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() ,input_auto[key].sum() ,delta=1e-2 ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Tuple = processor_wavaveca.batch_decode(A ) UpperCAmelCase__ : List[str] = processor_auto.batch_decode(A ) self.assertListEqual(decoded_wavaveca.text ,decoded_auto.text ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : int = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) self.assertListEqual( processor.model_input_names ,feature_extractor.model_input_names ,msg="""`processor` and `feature_extractor` model input names do not match""" ,) @staticmethod def __lowercase ( A : Optional[Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = self._get_dummy_logits()[0] UpperCAmelCase__ : List[str] = processor.decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ) ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""end_offset""" ) ,[1, 3, 5] ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : int = self._get_dummy_logits() UpperCAmelCase__ : Any = processor.batch_decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(A ,"""word""" ) ) for o in outputs["""word_offsets"""]] ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""end_offset""" ) ,[1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self : Tuple ): '''simple docstring''' import torch UpperCAmelCase__ : Any = load_dataset("""common_voice""" ,"""en""" ,split="""train""" ,streaming=A ) UpperCAmelCase__ : Tuple = ds.cast_column("""audio""" ,datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Tuple = iter(A ) UpperCAmelCase__ : Optional[int] = next(A ) UpperCAmelCase__ : List[Any] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase__ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Tuple = processor(sample["""audio"""]["""array"""] ,return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(A ).logits.cpu().numpy() UpperCAmelCase__ : Any = processor.decode(logits[0] ,output_word_offsets=A ) UpperCAmelCase__ : str = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase__ : Dict = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,A ) self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,output.text ) # output times UpperCAmelCase__ : str = torch.tensor(self.get_from_offsets(A ,"""start_time""" ) ) UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(A ,"""end_time""" ) ) # fmt: off UpperCAmelCase__ : Union[str, Any] = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : List[Any] = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) ) self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) )
65
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece.model') __UpperCAmelCase = {'target_lang': 'fi', 'source_lang': 'en'} __UpperCAmelCase = '>>zh<<' __UpperCAmelCase = 'Helsinki-NLP/' if is_torch_available(): __UpperCAmelCase = 'pt' elif is_tf_available(): __UpperCAmelCase = 'tf' else: __UpperCAmelCase = 'jax' @require_sentencepiece class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = MarianTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Optional[int] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Optional[Any] = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] UpperCAmelCase__ : int = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[int] = Path(self.tmpdirname ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(A ,save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(A ,save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) UpperCAmelCase__ : Dict = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self : List[Any] ,**A : List[Any] ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Union[str, Any] ,A : Tuple ): '''simple docstring''' return ( "This is a test", "This is a test", ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """</s>""" UpperCAmelCase__ : int = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""</s>""" ) self.assertEqual(vocab_keys[1] ,"""<unk>""" ) self.assertEqual(vocab_keys[-1] ,"""<pad>""" ) self.assertEqual(len(A ) ,9 ) def __lowercase ( self : Dict ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,9 ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MarianTokenizer.from_pretrained(f"{ORG_NAME}opus-mt-en-de" ) UpperCAmelCase__ : List[str] = en_de_tokenizer(["""I am a small frog"""] ,return_tensors=A ) self.assertIsInstance(A ,A ) UpperCAmelCase__ : str = [38, 121, 14, 697, 38_848, 0] self.assertListEqual(A ,batch.input_ids[0] ) UpperCAmelCase__ : Optional[Any] = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(A ) UpperCAmelCase__ : Tuple = [x.name for x in Path(A ).glob("""*""" )] self.assertIn("""source.spm""" ,A ) MarianTokenizer.from_pretrained(A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = tok( ["""I am a small frog""" * 1_000, """I am a small frog"""] ,padding=A ,truncation=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch.input_ids.shape ,(2, 512) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : Tuple = tok(["""I am a tiny frog""", """I am a small frog"""] ,padding=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch_smaller.input_ids.shape ,(2, 10) ) @slow def __lowercase ( self : Dict ): '''simple docstring''' # fmt: off UpperCAmelCase__ : Optional[int] = {"""input_ids""": [[43_495, 462, 20, 42_164, 1_369, 52, 464, 132, 1_703, 492, 13, 7_491, 38_999, 6, 8, 464, 132, 1_703, 492, 13, 4_669, 37_867, 13, 7_525, 27, 1_593, 988, 13, 33_972, 7_029, 6, 20, 8_251, 383, 2, 270, 5_866, 3_788, 2, 2_353, 8_251, 12_338, 2, 13_958, 387, 2, 3_629, 6_953, 188, 2_900, 2, 13_958, 8_011, 11_501, 23, 8_460, 4_073, 34_009, 20, 435, 11_439, 27, 8, 8_460, 4_073, 6_004, 20, 9_988, 375, 27, 33, 266, 1_945, 1_076, 1_350, 37_867, 3_288, 5, 577, 1_076, 4_374, 8, 5_082, 5, 26_453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10_767, 6, 316, 304, 4_239, 3, 0], [148, 15_722, 19, 1_839, 12, 1_350, 13, 22_327, 5_082, 5_418, 47_567, 35_938, 59, 318, 19_552, 108, 2_183, 54, 14_976, 4_835, 32, 547, 1_114, 8, 315, 2_417, 5, 92, 19_088, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100], [36, 6_395, 12_570, 39_147, 11_597, 6, 266, 4, 45_405, 7_296, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""Helsinki-NLP/opus-mt-en-de""" ,revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" ,decode_kwargs={"""use_source_tokenizer""": True} ,) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) UpperCAmelCase__ : Any = """Tämä on testi""" UpperCAmelCase__ : int = """This is a test""" UpperCAmelCase__ : List[str] = [76, 7, 2_047, 2] UpperCAmelCase__ : Optional[Any] = [69, 12, 11, 940, 2] UpperCAmelCase__ : List[str] = tokenizer(A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : Optional[int] = tokenizer(text_target=A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : int = tokenizer.decode(A ,skip_special_tokens=A ) self.assertEqual(A ,A )
65
1
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = '▁' __UpperCAmelCase = { 'vocab_file': 'vocab.json', 'spm_file': 'sentencepiece.bpe.model', 'tokenizer_config_file': 'tokenizer_config.json', } __UpperCAmelCase = { 'vocab_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json', }, 'spm_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_config_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json', }, } __UpperCAmelCase = { 'facebook/m2m100_418M': 1024, } # fmt: off __UpperCAmelCase = { 'm2m100': ['af', 'am', 'ar', 'ast', 'az', 'ba', 'be', 'bg', 'bn', 'br', 'bs', 'ca', 'ceb', 'cs', 'cy', 'da', 'de', 'el', 'en', 'es', 'et', 'fa', 'ff', 'fi', 'fr', 'fy', 'ga', 'gd', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hy', 'id', 'ig', 'ilo', 'is', 'it', 'ja', 'jv', 'ka', 'kk', 'km', 'kn', 'ko', 'lb', 'lg', 'ln', 'lo', 'lt', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'my', 'ne', 'nl', 'no', 'ns', 'oc', 'or', 'pa', 'pl', 'ps', 'pt', 'ro', 'ru', 'sd', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'ss', 'su', 'sv', 'sw', 'ta', 'th', 'tl', 'tn', 'tr', 'uk', 'ur', 'uz', 'vi', 'wo', 'xh', 'yi', 'yo', 'zh', 'zu'], 'wmt21': ['en', 'ha', 'is', 'ja', 'cs', 'ru', 'zh', 'de'] } class __lowercase ( __lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = ["""input_ids""", """attention_mask"""] snake_case_ = [] snake_case_ = [] def __init__( self : int ,A : List[Any] ,A : str ,A : List[Any]=None ,A : Dict=None ,A : str="<s>" ,A : int="</s>" ,A : List[Any]="</s>" ,A : Optional[Any]="<pad>" ,A : List[str]="<unk>" ,A : Optional[Any]="m2m100" ,A : Optional[Dict[str, Any]] = None ,A : List[str]=8 ,**A : Optional[Any] ,): '''simple docstring''' UpperCAmelCase__ : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs UpperCAmelCase__ : List[str] = language_codes UpperCAmelCase__ : Dict = FAIRSEQ_LANGUAGE_CODES[language_codes] UpperCAmelCase__ : Optional[int] = {lang_code: f"__{lang_code}__" for lang_code in fairseq_language_code} UpperCAmelCase__ : Tuple = kwargs.get("""additional_special_tokens""" ,[] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(A ) for lang_code in fairseq_language_code if self.get_lang_token(A ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=A ,tgt_lang=A ,bos_token=A ,eos_token=A ,sep_token=A ,unk_token=A ,pad_token=A ,language_codes=A ,sp_model_kwargs=self.sp_model_kwargs ,num_madeup_words=A ,**A ,) UpperCAmelCase__ : str = vocab_file UpperCAmelCase__ : Dict = load_json(A ) UpperCAmelCase__ : Optional[int] = {v: k for k, v in self.encoder.items()} UpperCAmelCase__ : Dict = spm_file UpperCAmelCase__ : List[Any] = load_spm(A ,self.sp_model_kwargs ) UpperCAmelCase__ : Dict = len(self.encoder ) UpperCAmelCase__ : Dict = { self.get_lang_token(A ): self.encoder_size + i for i, lang_code in enumerate(A ) } UpperCAmelCase__ : Union[str, Any] = {lang_code: self.encoder_size + i for i, lang_code in enumerate(A )} UpperCAmelCase__ : Union[str, Any] = {v: k for k, v in self.lang_token_to_id.items()} UpperCAmelCase__ : List[str] = src_lang if src_lang is not None else """en""" UpperCAmelCase__ : int = tgt_lang UpperCAmelCase__ : int = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) UpperCAmelCase__ : List[str] = num_madeup_words @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return len(self.encoder ) + len(self.lang_token_to_id ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self._src_lang @src_lang.setter def __lowercase ( self : List[str] ,A : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __lowercase ( self : Optional[Any] ,A : str ): '''simple docstring''' return self.sp_model.encode(A ,out_type=A ) def __lowercase ( self : List[str] ,A : Dict ): '''simple docstring''' if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(A ,self.encoder[self.unk_token] ) def __lowercase ( self : Any ,A : int ): '''simple docstring''' if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(A ,self.unk_token ) def __lowercase ( self : Tuple ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = [] UpperCAmelCase__ : str = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(A ) + token UpperCAmelCase__ : str = [] else: current_sub_tokens.append(A ) out_string += self.sp_model.decode(A ) return out_string.strip() def __lowercase ( self : Optional[int] ,A : List[int] ,A : Optional[List[int]] = None ,A : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A ,token_ids_a=A ,already_has_special_tokens=A ) UpperCAmelCase__ : Union[str, Any] = [1] * len(self.prefix_tokens ) UpperCAmelCase__ : Optional[int] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(A )) + suffix_ones return prefix_ones + ([0] * len(A )) + ([0] * len(A )) + suffix_ones def __lowercase ( self : Tuple ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = {self.convert_ids_to_tokens(A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.__dict__.copy() UpperCAmelCase__ : Dict = None return state def __setstate__( self : int ,A : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : List[Any] = load_spm(self.spm_file ,self.sp_model_kwargs ) def __lowercase ( self : Any ,A : str ,A : Optional[str] = None ): '''simple docstring''' UpperCAmelCase__ : List[str] = Path(A ) if not save_dir.is_dir(): raise OSError(f"{save_directory} should be a directory" ) UpperCAmelCase__ : Tuple = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""vocab_file"""] ) UpperCAmelCase__ : Any = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""spm_file"""] ) save_json(self.encoder ,A ) if os.path.abspath(self.spm_file ) != os.path.abspath(A ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file ,A ) elif not os.path.isfile(self.spm_file ): with open(A ,"""wb""" ) as fi: UpperCAmelCase__ : List[str] = self.sp_model.serialized_model_proto() fi.write(A ) return (str(A ), str(A )) def __lowercase ( self : str ,A : List[str] ,A : str = "en" ,A : Optional[List[str]] = None ,A : str = "ro" ,**A : List[Any] ,): '''simple docstring''' UpperCAmelCase__ : List[Any] = src_lang UpperCAmelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(A ,A ,**A ) def __lowercase ( self : Any ,A : Union[str, Any] ,A : Optional[str] ,A : Optional[str] ,**A : List[Any] ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) UpperCAmelCase__ : List[Any] = src_lang UpperCAmelCase__ : List[str] = self(A ,add_special_tokens=A ,**A ) UpperCAmelCase__ : List[Any] = self.get_lang_id(A ) UpperCAmelCase__ : List[str] = tgt_lang_id return inputs def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self.set_src_lang_special_tokens(self.src_lang ) def __lowercase ( self : List[Any] ): '''simple docstring''' self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowercase ( self : int ,A : str ): '''simple docstring''' UpperCAmelCase__ : int = self.get_lang_token(A ) UpperCAmelCase__ : List[str] = self.lang_token_to_id[lang_token] UpperCAmelCase__ : Union[str, Any] = [self.cur_lang_id] UpperCAmelCase__ : int = [self.eos_token_id] def __lowercase ( self : Dict ,A : str ): '''simple docstring''' UpperCAmelCase__ : int = self.get_lang_token(A ) UpperCAmelCase__ : List[Any] = self.lang_token_to_id[lang_token] UpperCAmelCase__ : Optional[int] = [self.cur_lang_id] UpperCAmelCase__ : str = [self.eos_token_id] def __lowercase ( self : int ,A : str ): '''simple docstring''' return self.lang_code_to_token[lang] def __lowercase ( self : Dict ,A : str ): '''simple docstring''' UpperCAmelCase__ : int = self.get_lang_token(A ) return self.lang_token_to_id[lang_token] def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = sentencepiece.SentencePieceProcessor(**__UpperCamelCase ) spm.Load(str(__UpperCamelCase ) ) return spm def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """r""" ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=2 )
65
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=__lowerCamelCase ): snake_case_ = ["""onnx"""] def __init__( self : int ,*A : List[str] ,**A : int ): '''simple docstring''' requires_backends(self ,["""onnx"""] ) @classmethod def __lowercase ( cls : Optional[Any] ,*A : List[str] ,**A : Dict ): '''simple docstring''' requires_backends(cls ,["""onnx"""] ) @classmethod def __lowercase ( cls : List[Any] ,*A : Optional[int] ,**A : int ): '''simple docstring''' requires_backends(cls ,["""onnx"""] )
65
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): def __init__( self : List[str] ,*A : Union[str, Any] ,**A : Tuple ): '''simple docstring''' warnings.warn( """The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use DPTImageProcessor instead.""" ,A ,) super().__init__(*A ,**A )
65
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) UpperCAmelCase__ : List[Any] = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(__UpperCamelCase ) # Let's go UpperCAmelCase__ : int = parser.parse_args() if not hasattr(__UpperCamelCase , """func""" ): parser.print_help() exit(1 ) # Run UpperCAmelCase__ : Union[str, Any] = args.func(__UpperCamelCase ) service.run() if __name__ == "__main__": main()
65
1
"""simple docstring""" __UpperCAmelCase = range(2, 20 + 1) __UpperCAmelCase = [10**k for k in range(ks[-1] + 1)] __UpperCAmelCase = {} def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = sum(a_i[j] for j in range(__UpperCamelCase , len(__UpperCamelCase ) ) ) UpperCAmelCase__ : str = sum(a_i[j] * base[j] for j in range(min(len(__UpperCamelCase ) , __UpperCamelCase ) ) ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = 0, 0 UpperCAmelCase__ : List[str] = n - i UpperCAmelCase__ : Tuple = memo.get(__UpperCamelCase ) if sub_memo is not None: UpperCAmelCase__ : List[Any] = sub_memo.get(__UpperCamelCase ) if jumps is not None and len(__UpperCamelCase ) > 0: # find and make the largest jump without going over UpperCAmelCase__ : str = -1 for _k in range(len(__UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCAmelCase__ : Tuple = _k break if max_jump >= 0: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c UpperCAmelCase__ : Optional[int] = diff + c for j in range(min(__UpperCamelCase , len(__UpperCamelCase ) ) ): UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = divmod(__UpperCamelCase , 10 ) if new_c > 0: add(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: UpperCAmelCase__ : Optional[Any] = [] else: UpperCAmelCase__ : Optional[int] = {c: []} UpperCAmelCase__ : Union[str, Any] = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCAmelCase__ , UpperCAmelCase__ : List[str] = next_term(__UpperCamelCase , k - 1 , i + dn , __UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCAmelCase__ , UpperCAmelCase__ : Tuple = compute(__UpperCamelCase , __UpperCamelCase , i + dn , __UpperCamelCase ) diff += _diff dn += terms_jumped UpperCAmelCase__ : Any = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCAmelCase__ : str = 0 while j < len(__UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(__UpperCamelCase , (diff, dn, k) ) return (diff, dn) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if i >= n: return 0, i if k > len(__UpperCamelCase ): a_i.extend([0 for _ in range(k - len(__UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCAmelCase__ : Optional[int] = i UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Dict = 0, 0, 0 for j in range(len(__UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCAmelCase__ : str = ds_c + ds_b diff += addend UpperCAmelCase__ : Dict = 0 for j in range(__UpperCamelCase ): UpperCAmelCase__ : Optional[Any] = a_i[j] + addend UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = divmod(__UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return diff, i - start_i def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' for j in range(__UpperCamelCase , len(__UpperCamelCase ) ): UpperCAmelCase__ : Tuple = digits[j] + addend if s >= 10: UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = divmod(__UpperCamelCase , 10 ) UpperCAmelCase__ : int = addend // 10 + quotient else: UpperCAmelCase__ : int = s UpperCAmelCase__ : Tuple = addend // 10 if addend == 0: break while addend > 0: UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = divmod(__UpperCamelCase , 10 ) digits.append(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase = 10**15 ): '''simple docstring''' UpperCAmelCase__ : Any = [1] UpperCAmelCase__ : List[str] = 1 UpperCAmelCase__ : Tuple = 0 while True: UpperCAmelCase__ , UpperCAmelCase__ : Any = next_term(__UpperCamelCase , 20 , i + dn , __UpperCamelCase ) dn += terms_jumped if dn == n - i: break UpperCAmelCase__ : Any = 0 for j in range(len(__UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F"{solution() = }")
65
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __UpperCAmelCase = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __lowercase : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : List[Any] ,A : int ,A : Optional[Any]=13 ,A : Dict=7 ,A : Dict=True ,A : Any=False ,A : Dict=99 ,A : int=32 ,A : Optional[int]=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : str=0.1 ,A : int=0.1 ,A : Optional[int]=20 ,A : Tuple=2 ,A : str=1 ,A : Optional[Any]=0 ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : int = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Union[str, Any] = pad_token_id UpperCAmelCase__ : List[str] = bos_token_id def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] ,self.vocab_size ).clip(3 ,self.vocab_size ) UpperCAmelCase__ : List[str] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) ,1 ) UpperCAmelCase__ : Any = np.concatenate([input_ids, eos_tensor] ,axis=1 ) UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : str = self.config_cls( vocab_size=self.vocab_size ,d_model=self.hidden_size ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,eos_token_ids=[2] ,bos_token_id=self.bos_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.pad_token_id ,**self.config_updates ,) UpperCAmelCase__ : Optional[Any] = prepare_pegasus_inputs_dict(A ,A ,A ) return config, inputs_dict def __lowercase ( self : Any ,A : Optional[int] ,A : str ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : Dict = model_class_name(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) ,dtype="""i4""" ) UpperCAmelCase__ : str = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Optional[int] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : int = model.decode( decoder_input_ids[:, -1:] ,A ,decoder_attention_mask=A ,past_key_values=outputs_cache.past_key_values ,decoder_position_ids=A ,) UpperCAmelCase__ : Dict = model.decode(A ,A ) UpperCAmelCase__ : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def __lowercase ( self : Optional[int] ,A : str ,A : Dict ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : str = model_class_name(A ) UpperCAmelCase__ : Any = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] ,axis=-1 ,) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : Dict = model.decode( decoder_input_ids[:, -1:] ,A ,past_key_values=outputs_cache.past_key_values ,decoder_attention_mask=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Union[str, Any] = model.decode(A ,A ,decoder_attention_mask=A ) UpperCAmelCase__ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , ): '''simple docstring''' if attention_mask is None: UpperCAmelCase__ : Union[str, Any] = np.not_equal(__UpperCamelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Tuple = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) snake_case_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = FlaxPegasusModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ) def __lowercase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(A ,A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(A ,A ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : List[Any] = self._prepare_for_class(A ,A ) UpperCAmelCase__ : int = model_class(A ) @jax.jit def encode_jitted(A : Optional[int] ,A : Union[str, Any]=None ,**A : Optional[Any] ): return model.encode(input_ids=A ,attention_mask=A ) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : int = encode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : Dict = encode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : Dict = model_class(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ,inputs_dict["""attention_mask"""] ) UpperCAmelCase__ : Dict = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(A : List[Any] ,A : Any ,A : List[Any] ): return model.decode( decoder_input_ids=A ,decoder_attention_mask=A ,encoder_outputs=A ,) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : Tuple = decode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : str = decode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""google/pegasus-large""" ,from_pt=A ) UpperCAmelCase__ : Any = np.ones((1, 1) ) UpperCAmelCase__ : Optional[Any] = model(A ) self.assertIsNotNone(A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Optional[Any] = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Union[str, Any] = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] UpperCAmelCase__ : str = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] UpperCAmelCase__ : str = tokenizer(A ,return_tensors="""np""" ,truncation=A ,max_length=512 ,padding=A ) UpperCAmelCase__ : Union[str, Any] = model.generate(**A ,num_beams=2 ).sequences UpperCAmelCase__ : int = tokenizer.batch_decode(A ,skip_special_tokens=A ) assert tgt_text == decoded
65
1
"""simple docstring""" from __future__ import annotations def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): '''simple docstring''' if (stress, tangential_force, area).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif stress < 0: raise ValueError("""Stress cannot be negative""" ) elif tangential_force < 0: raise ValueError("""Tangential Force cannot be negative""" ) elif area < 0: raise ValueError("""Area cannot be negative""" ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
65
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) UpperCAmelCase__ : Union[str, Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowercase ( __lowerCamelCase ): snake_case_ = ["""image_processor""", """tokenizer"""] snake_case_ = """ChineseCLIPImageProcessor""" snake_case_ = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self : int ,A : Union[str, Any]=None ,A : Optional[Any]=None ,**A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase__ : Optional[Any] = kwargs.pop("""feature_extractor""" ) UpperCAmelCase__ : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) UpperCAmelCase__ : List[str] = self.image_processor def __call__( self : int ,A : Optional[int]=None ,A : Union[str, Any]=None ,A : str=None ,**A : Union[str, Any] ): '''simple docstring''' if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase__ : str = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase__ : Optional[int] = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase__ : List[str] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def __lowercase ( self : Tuple ,*A : int ,**A : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*A ,**A ) def __lowercase ( self : str ,*A : Optional[int] ,**A : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*A ,**A ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Any = self.tokenizer.model_input_names UpperCAmelCase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def __lowercase ( self : int ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class
65
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase ( __UpperCamelCase = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCAmelCase__ : Dict = F"{olid} is not a valid Open Library olid" raise ValueError(__UpperCamelCase ) return requests.get(F"https://openlibrary.org/{new_olid}.json" ).json() def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } UpperCAmelCase__ : Dict = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase__ : str = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] UpperCAmelCase__ : Dict = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Dict = """, """.join(__UpperCamelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __UpperCAmelCase = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: __UpperCAmelCase = summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('\n'.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
65
1
"""simple docstring""" import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } __UpperCAmelCase = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } __UpperCAmelCase = {'facebook/blenderbot_small-90M': 512} def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = set() UpperCAmelCase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ : Union[str, Any] = char UpperCAmelCase__ : List[Any] = set(__UpperCamelCase ) return pairs class __lowercase ( __lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[int] ,A : Tuple ,A : int ,A : int="__start__" ,A : Dict="__end__" ,A : int="__unk__" ,A : List[Any]="__null__" ,**A : Optional[int] ,): '''simple docstring''' super().__init__(unk_token=A ,bos_token=A ,eos_token=A ,pad_token=A ,**A ) with open(A ,encoding="""utf-8""" ) as vocab_handle: UpperCAmelCase__ : Dict = json.load(A ) UpperCAmelCase__ : Optional[int] = {v: k for k, v in self.encoder.items()} with open(A ,encoding="""utf-8""" ) as merges_handle: UpperCAmelCase__ : Any = merges_handle.read().split("""\n""" )[1:-1] UpperCAmelCase__ : Any = [tuple(merge.split() ) for merge in merges] UpperCAmelCase__ : List[str] = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : str = {} @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return len(self.encoder ) def __lowercase ( self : Any ): '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def __lowercase ( self : str ,A : str ): '''simple docstring''' if token in self.cache: return self.cache[token] UpperCAmelCase__ : List[str] = re.sub("""([.,!?()])""" ,R""" \1""" ,A ) UpperCAmelCase__ : Any = re.sub("""(')""" ,R""" \1 """ ,A ) UpperCAmelCase__ : List[Any] = re.sub(R"""\s{2,}""" ,""" """ ,A ) if "\n" in token: UpperCAmelCase__ : str = token.replace("""\n""" ,""" __newln__""" ) UpperCAmelCase__ : int = token.split(""" """ ) UpperCAmelCase__ : int = [] for token in tokens: if not len(A ): continue UpperCAmelCase__ : Union[str, Any] = token.lower() UpperCAmelCase__ : List[str] = tuple(A ) UpperCAmelCase__ : List[Any] = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) UpperCAmelCase__ : List[Any] = get_pairs(A ) if not pairs: words.append(A ) continue while True: UpperCAmelCase__ : int = min(A ,key=lambda A : self.bpe_ranks.get(A ,float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ : Tuple = bigram UpperCAmelCase__ : str = [] UpperCAmelCase__ : List[str] = 0 while i < len(A ): try: UpperCAmelCase__ : Tuple = word.index(A ,A ) new_word.extend(word[i:j] ) UpperCAmelCase__ : Optional[Any] = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ : Optional[int] = tuple(A ) UpperCAmelCase__ : Tuple = new_word if len(A ) == 1: break else: UpperCAmelCase__ : Optional[int] = get_pairs(A ) UpperCAmelCase__ : List[str] = """@@ """.join(A ) UpperCAmelCase__ : int = word[:-4] UpperCAmelCase__ : List[str] = word words.append(A ) return " ".join(A ) def __lowercase ( self : str ,A : str ): '''simple docstring''' UpperCAmelCase__ : Dict = [] UpperCAmelCase__ : Dict = re.findall(R"""\S+\n?""" ,A ) for token in words: split_tokens.extend(list(self.bpe(A ).split(""" """ ) ) ) return split_tokens def __lowercase ( self : Optional[Any] ,A : str ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = token.lower() return self.encoder.get(A ,self.encoder.get(self.unk_token ) ) def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' return self.decoder.get(A ,self.unk_token ) def __lowercase ( self : Union[str, Any] ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = """ """.join(A ).replace("""@@ """ ,"""""" ).strip() return out_string def __lowercase ( self : Tuple ,A : str ,A : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(A ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase__ : int = os.path.join( A ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Optional[int] = os.path.join( A ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(A ,"""w""" ,encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=A ,ensure_ascii=A ) + """\n""" ) UpperCAmelCase__ : Union[str, Any] = 0 with open(A ,"""w""" ,encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda A : kv[1] ): if index != token_index: logger.warning( f"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." """ Please check that the tokenizer is not corrupted!""" ) UpperCAmelCase__ : Dict = token_index writer.write(""" """.join(A ) + """\n""" ) index += 1 return vocab_file, merge_file
65
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=True , __UpperCamelCase="pt" ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {"""add_prefix_space""": True} if isinstance(__UpperCamelCase , __UpperCamelCase ) and not line.startswith(""" """ ) else {} UpperCAmelCase__ : List[str] = padding_side return tokenizer( [line] , max_length=__UpperCamelCase , padding="""max_length""" if pad_to_max_length else None , truncation=__UpperCamelCase , return_tensors=__UpperCamelCase , add_special_tokens=__UpperCamelCase , **__UpperCamelCase , ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , ): '''simple docstring''' UpperCAmelCase__ : str = input_ids.ne(__UpperCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ,A : List[Any] ,A : Union[str, Any] ,A : Any ,A : Optional[int] ,A : Union[str, Any]="train" ,A : Tuple=None ,A : Union[str, Any]=None ,A : Tuple=None ,A : int="" ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : Optional[Any] = Path(A ).joinpath(type_path + """.source""" ) UpperCAmelCase__ : List[str] = Path(A ).joinpath(type_path + """.target""" ) UpperCAmelCase__ : Dict = self.get_char_lens(self.src_file ) UpperCAmelCase__ : int = max_source_length UpperCAmelCase__ : List[str] = max_target_length assert min(self.src_lens ) > 0, f"found empty line in {self.src_file}" UpperCAmelCase__ : Dict = tokenizer UpperCAmelCase__ : str = prefix if n_obs is not None: UpperCAmelCase__ : int = self.src_lens[:n_obs] UpperCAmelCase__ : Any = src_lang UpperCAmelCase__ : Any = tgt_lang def __len__( self : Optional[Any] ): '''simple docstring''' return len(self.src_lens ) def __getitem__( self : Union[str, Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = index + 1 # linecache starts at 1 UpperCAmelCase__ : Tuple = self.prefix + linecache.getline(str(self.src_file ) ,A ).rstrip("""\n""" ) UpperCAmelCase__ : Dict = linecache.getline(str(self.tgt_file ) ,A ).rstrip("""\n""" ) assert source_line, f"empty source line for index {index}" assert tgt_line, f"empty tgt line for index {index}" # Need to add eos token manually for T5 if isinstance(self.tokenizer ,A ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right UpperCAmelCase__ : str = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,A ) else self.tokenizer ) UpperCAmelCase__ : Tuple = self.tokenizer.generator if isinstance(self.tokenizer ,A ) else self.tokenizer UpperCAmelCase__ : Tuple = encode_line(A ,A ,self.max_source_length ,"""right""" ) UpperCAmelCase__ : Dict = encode_line(A ,A ,self.max_target_length ,"""right""" ) UpperCAmelCase__ : Optional[Any] = source_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : List[str] = target_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : Union[str, Any] = source_inputs["""attention_mask"""].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def __lowercase ( A : int ): '''simple docstring''' return [len(A ) for x in Path(A ).open().readlines()] def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : int = torch.stack([x["""input_ids"""] for x in batch] ) UpperCAmelCase__ : Union[str, Any] = torch.stack([x["""attention_mask"""] for x in batch] ) UpperCAmelCase__ : Any = torch.stack([x["""decoder_input_ids"""] for x in batch] ) UpperCAmelCase__ : List[Any] = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : Any = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : str = trim_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = trim_batch(A ,A ,attention_mask=A ) UpperCAmelCase__ : List[str] = { """input_ids""": source_ids, """attention_mask""": source_mask, """decoder_input_ids""": y, } return batch __UpperCAmelCase = getLogger(__name__) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return list(itertools.chain.from_iterable(__UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = get_git_info() save_json(__UpperCamelCase , os.path.join(__UpperCamelCase , """git_log.json""" ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=4 , **__UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=__UpperCamelCase , **__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = git.Repo(search_parent_directories=__UpperCamelCase ) UpperCAmelCase__ : List[str] = { """repo_id""": str(__UpperCamelCase ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), """hostname""": str(socket.gethostname() ), } return repo_infos def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return list(map(__UpperCamelCase , __UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """wb""" ) as f: return pickle.dump(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' def remove_articles(__UpperCamelCase ): return re.sub(r"""\b(a|an|the)\b""" , """ """ , __UpperCamelCase ) def white_space_fix(__UpperCamelCase ): return " ".join(text.split() ) def remove_punc(__UpperCamelCase ): UpperCAmelCase__ : List[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCamelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCamelCase ) ) ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : Dict = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : int = Counter(__UpperCamelCase ) & Counter(__UpperCamelCase ) UpperCAmelCase__ : List[str] = sum(common.values() ) if num_same == 0: return 0 UpperCAmelCase__ : str = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Tuple = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return normalize_answer(__UpperCamelCase ) == normalize_answer(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' assert len(__UpperCamelCase ) == len(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 0 for hypo, pred in zip(__UpperCamelCase , __UpperCamelCase ): em += exact_match_score(__UpperCamelCase , __UpperCamelCase ) if len(__UpperCamelCase ) > 0: em /= len(__UpperCamelCase ) return {"em": em} def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return model_prefix.startswith("""rag""" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead UpperCAmelCase__ : str = """dropout_rate""" for p in extra_params: if getattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): if not hasattr(__UpperCamelCase , __UpperCamelCase ) and not hasattr(__UpperCamelCase , equivalent_param[p] ): logger.info("""config doesn't have a `{}` attribute""".format(__UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) continue UpperCAmelCase__ : Tuple = p if hasattr(__UpperCamelCase , __UpperCamelCase ) else equivalent_param[p] setattr(__UpperCamelCase , __UpperCamelCase , getattr(__UpperCamelCase , __UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) return hparams, config
65
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase = { 'configuration_convbert': ['CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConvBertConfig', 'ConvBertOnnxConfig'], 'tokenization_convbert': ['ConvBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['ConvBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'ConvBertForMaskedLM', 'ConvBertForMultipleChoice', 'ConvBertForQuestionAnswering', 'ConvBertForSequenceClassification', 'ConvBertForTokenClassification', 'ConvBertLayer', 'ConvBertModel', 'ConvBertPreTrainedModel', 'load_tf_weights_in_convbert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFConvBertForMaskedLM', 'TFConvBertForMultipleChoice', 'TFConvBertForQuestionAnswering', 'TFConvBertForSequenceClassification', 'TFConvBertForTokenClassification', 'TFConvBertLayer', 'TFConvBertModel', 'TFConvBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaControlnetPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return 32 @property def __lowercase ( self : int ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self : Any ): '''simple docstring''' return 100 @property def __lowercase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Tuple = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } UpperCAmelCase__ : int = UNetaDConditionModel(**A ) return model @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : str = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : str = self.dummy_unet UpperCAmelCase__ : List[Any] = self.dummy_movq UpperCAmelCase__ : List[Any] = DDIMScheduler( num_train_timesteps=1_000 ,beta_schedule="""linear""" ,beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,clip_sample=A ,set_alpha_to_one=A ,steps_offset=1 ,prediction_type="""epsilon""" ,thresholding=A ,) UpperCAmelCase__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __lowercase ( self : str ,A : Optional[Any] ,A : Any=0 ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( A ) # create hint UpperCAmelCase__ : int = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) if str(A ).startswith("""mps""" ): UpperCAmelCase__ : Optional[int] = torch.manual_seed(A ) else: UpperCAmelCase__ : Dict = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase__ : Dict = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = """cpu""" UpperCAmelCase__ : List[Any] = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = self.pipeline_class(**A ) UpperCAmelCase__ : Optional[int] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = pipe(**self.get_dummy_inputs(A ) ) UpperCAmelCase__ : Tuple = output.images UpperCAmelCase__ : Dict = pipe( **self.get_dummy_inputs(A ) ,return_dict=A ,)[0] UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1] UpperCAmelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Optional[int] = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : Union[str, Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) UpperCAmelCase__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) UpperCAmelCase__ : int = torch.from_numpy(np.array(A ) ).float() / 2_5_5.0 UpperCAmelCase__ : Union[str, Any] = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) UpperCAmelCase__ : List[str] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" ,torch_dtype=torch.floataa ) pipe_prior.to(A ) UpperCAmelCase__ : List[Any] = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" ,torch_dtype=torch.floataa ) UpperCAmelCase__ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[Any] = """A robot, 4k photo""" UpperCAmelCase__ : List[Any] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = pipe_prior( A ,generator=A ,num_inference_steps=5 ,negative_prompt="""""" ,).to_tuple() UpperCAmelCase__ : List[str] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ : int = pipeline( image_embeds=A ,negative_image_embeds=A ,hint=A ,generator=A ,num_inference_steps=100 ,output_type="""np""" ,) UpperCAmelCase__ : Any = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A ,A )
65
1
"""simple docstring""" import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class __lowercase : def __init__( self : List[str] ,A : int ,A : str=3 ,A : str=7 ,A : Tuple=True ,A : str=True ,A : Optional[int]=False ,A : Dict=True ,A : List[str]=99 ,A : Optional[int]=32 ,A : List[Any]=5 ,A : Optional[int]=4 ,A : Union[str, Any]=37 ,A : Optional[int]="gelu" ,A : Optional[Any]=0.1 ,A : Tuple=0.1 ,A : Optional[int]=512 ,A : Optional[Any]=16 ,A : List[str]=2 ,A : str=0.0_2 ,A : Optional[Any]=3 ,A : Optional[Any]=4 ,A : Optional[Any]=None ,): '''simple docstring''' UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : str = batch_size UpperCAmelCase__ : Optional[int] = seq_length UpperCAmelCase__ : Optional[Any] = is_training UpperCAmelCase__ : Optional[Any] = use_input_mask UpperCAmelCase__ : List[Any] = use_token_type_ids UpperCAmelCase__ : Union[str, Any] = use_labels UpperCAmelCase__ : Tuple = vocab_size UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Union[str, Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : Optional[Any] = attention_probs_dropout_prob UpperCAmelCase__ : Union[str, Any] = max_position_embeddings UpperCAmelCase__ : Optional[Any] = type_vocab_size UpperCAmelCase__ : int = type_sequence_label_size UpperCAmelCase__ : str = initializer_range UpperCAmelCase__ : List[str] = num_labels UpperCAmelCase__ : Union[str, Any] = num_choices UpperCAmelCase__ : Dict = scope def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : Any = None if self.use_input_mask: UpperCAmelCase__ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Dict = None UpperCAmelCase__ : int = None UpperCAmelCase__ : List[Any] = None UpperCAmelCase__ : Optional[int] = None if self.use_labels: UpperCAmelCase__ : Any = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase__ : str = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase__ : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return FalconConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A ,initializer_range=self.initializer_range ,pad_token_id=1 ,new_decoder_architecture=A ,) def __lowercase ( self : Any ,A : Optional[Any] ,A : Optional[Any] ,A : Union[str, Any] ,A : Union[str, Any] ,A : Optional[Any] ,A : Any ,A : Dict ): '''simple docstring''' UpperCAmelCase__ : str = FalconModel(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : int = model(A ,attention_mask=A ) UpperCAmelCase__ : List[str] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def __lowercase ( self : Union[str, Any] ,A : str ,A : str ,A : Optional[int] ,A : List[str] ,A : List[str] ,A : Dict ,A : List[str] ,A : Tuple ,A : List[str] ,): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : Optional[int] = FalconModel(A ) model.to(A ) model.eval() UpperCAmelCase__ : str = model( A ,attention_mask=A ,encoder_hidden_states=A ,encoder_attention_mask=A ,) UpperCAmelCase__ : List[Any] = model( A ,attention_mask=A ,encoder_hidden_states=A ,) UpperCAmelCase__ : int = model(A ,attention_mask=A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def __lowercase ( self : Dict ,A : Tuple ,A : Dict ,A : Union[str, Any] ,A : Any ,A : Union[str, Any] ,A : Dict ,A : int ,A : Tuple ,A : str ,): '''simple docstring''' UpperCAmelCase__ : List[str] = FalconForCausalLM(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : Tuple = model(A ,attention_mask=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def __lowercase ( self : str ,A : Any ,A : Optional[Any] ,A : str ,A : Any ,A : List[str] ,A : Any ,A : Dict ,A : str ,A : Any ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : int = True UpperCAmelCase__ : List[str] = FalconForCausalLM(config=A ) model.to(A ) model.eval() # first forward pass UpperCAmelCase__ : List[str] = model( A ,attention_mask=A ,encoder_hidden_states=A ,encoder_attention_mask=A ,use_cache=A ,) UpperCAmelCase__ : int = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase__ : str = ids_tensor((self.batch_size, 3) ,config.vocab_size ) UpperCAmelCase__ : Optional[Any] = ids_tensor((self.batch_size, 3) ,vocab_size=2 ) # append to next input_ids and UpperCAmelCase__ : Optional[Any] = torch.cat([input_ids, next_tokens] ,dim=-1 ) UpperCAmelCase__ : List[str] = torch.cat([input_mask, next_mask] ,dim=-1 ) UpperCAmelCase__ : Dict = model( A ,attention_mask=A ,encoder_hidden_states=A ,encoder_attention_mask=A ,output_hidden_states=A ,)["""hidden_states"""][0] UpperCAmelCase__ : Any = model( A ,attention_mask=A ,encoder_hidden_states=A ,encoder_attention_mask=A ,past_key_values=A ,output_hidden_states=A ,)["""hidden_states"""][0] # select random slice UpperCAmelCase__ : List[str] = ids_tensor((1,) ,output_from_past.shape[-1] ).item() UpperCAmelCase__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase__ : List[str] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A ,A ,atol=1e-3 ) ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Tuple = config_and_inputs UpperCAmelCase__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __lowercase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) snake_case_ = (FalconForCausalLM,) if is_torch_available() else () snake_case_ = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = FalconModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ,hidden_size=37 ) def __lowercase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , *UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: UpperCAmelCase__ : List[Any] = alibi self.model_tester.create_and_check_model(A ,*A ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Tuple = 3 UpperCAmelCase__ : Dict = input_dict["""input_ids"""] UpperCAmelCase__ : Optional[int] = input_ids.ne(1 ).to(A ) UpperCAmelCase__ : Tuple = ids_tensor([self.model_tester.batch_size] ,self.model_tester.type_sequence_label_size ) UpperCAmelCase__ : str = FalconForSequenceClassification(A ) model.to(A ) model.eval() UpperCAmelCase__ : Optional[int] = model(A ,attention_mask=A ,labels=A ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Dict = 3 UpperCAmelCase__ : List[str] = """single_label_classification""" UpperCAmelCase__ : Optional[int] = input_dict["""input_ids"""] UpperCAmelCase__ : Tuple = input_ids.ne(1 ).to(A ) UpperCAmelCase__ : Optional[Any] = ids_tensor([self.model_tester.batch_size] ,self.model_tester.type_sequence_label_size ) UpperCAmelCase__ : int = FalconForSequenceClassification(A ) model.to(A ) model.eval() UpperCAmelCase__ : List[str] = model(A ,attention_mask=A ,labels=A ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Union[str, Any] = input_dict["""input_ids"""] UpperCAmelCase__ : Optional[int] = FalconForCausalLM(A ) model.to(A ) model.eval() UpperCAmelCase__ : Optional[Any] = model(A ,use_cache=A ) UpperCAmelCase__ : Any = input_ids.shape[0] UpperCAmelCase__ : Optional[Any] = model._convert_to_rw_cache(result.past_key_values ) UpperCAmelCase__ : int = model._convert_cache_to_standard_format(A ,A ) for layer in range(len(A ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Union[str, Any] = 3 UpperCAmelCase__ : Optional[Any] = """multi_label_classification""" UpperCAmelCase__ : Tuple = input_dict["""input_ids"""] UpperCAmelCase__ : Union[str, Any] = input_ids.ne(1 ).to(A ) UpperCAmelCase__ : Tuple = ids_tensor( [self.model_tester.batch_size, config.num_labels] ,self.model_tester.type_sequence_label_size ).to(torch.float ) UpperCAmelCase__ : Tuple = FalconForSequenceClassification(A ) model.to(A ) model.eval() UpperCAmelCase__ : int = model(A ,attention_mask=A ,labels=A ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' # Falcon can have different numbers of KV-heads than the number of query heads, so we need # to override this test to use the right head counts. for model_class in self.all_generative_model_classes: UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(A ,"""use_cache""" ): return UpperCAmelCase__ : List[Any] = model_class(A ).to(A ) if "use_cache" not in inputs: UpperCAmelCase__ : Optional[int] = True UpperCAmelCase__ : List[Any] = model(**A ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return UpperCAmelCase__ : int = ( getattr(A ,"""decoder_layers""" ,A ) or getattr(A ,"""num_decoder_layers""" ,A ) or config.num_hidden_layers ) UpperCAmelCase__ : Optional[Any] = getattr(A ,"""num_kv_heads""" ,config.num_attention_heads ) UpperCAmelCase__ : Dict = getattr(A ,"""d_model""" ,config.hidden_size ) UpperCAmelCase__ : Optional[int] = embed_dim // num_attention_heads UpperCAmelCase__ : str = outputs["""past_key_values"""] self.assertEqual(len(A ) ,A ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = inputs["""input_ids"""].shape for i in range(A ): if config.new_decoder_architecture: UpperCAmelCase__ : List[Any] = config.num_attention_heads elif config.multi_query: UpperCAmelCase__ : Dict = 1 self.assertEqual(len(past_kv[0] ) ,2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape ,(batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape ,(batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : str = AutoTokenizer.from_pretrained("""Rocketknight1/falcon-rw-1b""" ) UpperCAmelCase__ : str = FalconForCausalLM.from_pretrained("""Rocketknight1/falcon-rw-1b""" ) model.eval() model.to(A ) UpperCAmelCase__ : Optional[int] = tokenizer("""My favorite food is""" ,return_tensors="""pt""" ).to(A ) UpperCAmelCase__ : Tuple = ( """My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday.""" ) UpperCAmelCase__ : Any = model.generate(**A ,do_sample=A ,max_new_tokens=19 ) UpperCAmelCase__ : Union[str, Any] = tokenizer.batch_decode(A )[0] self.assertEqual(A ,A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' # The big models are way too big for the CI, so we use tiny random models that resemble their # architectures but with much smaller and fewer layers for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: UpperCAmelCase__ : Any = AutoTokenizer.from_pretrained(A ) UpperCAmelCase__ : Dict = FalconForCausalLM.from_pretrained(A ) model.eval() model.to(A ) UpperCAmelCase__ : Dict = tokenizer("""My favorite food is""" ,return_tensors="""pt""" ).to(A ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**A ,do_sample=A ,max_new_tokens=4 ) model.generate(**A ,do_sample=A ,max_new_tokens=4 ) model.generate(**A ,num_beams=2 ,max_new_tokens=4 ) @slow def __lowercase ( self : str ): '''simple docstring''' # The big models are way too big for the CI, so we use tiny random models that resemble their # architectures but with much smaller and fewer layers with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: UpperCAmelCase__ : List[Any] = AutoTokenizer.from_pretrained(A ) UpperCAmelCase__ : str = FalconForCausalLM.from_pretrained(A ) model.eval() model.to(device=A ) UpperCAmelCase__ : Optional[Any] = tokenizer("""My favorite food is""" ,return_tensors="""pt""" ).to(A ) # Test results are the same with and without cache UpperCAmelCase__ : Optional[Any] = model.generate(**A ,do_sample=A ,max_new_tokens=20 ,use_cache=A ) UpperCAmelCase__ : Optional[Any] = model.generate(**A ,do_sample=A ,max_new_tokens=20 ,use_cache=A ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
65
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = """vision-encoder-decoder""" snake_case_ = True def __init__( self : List[Any] ,**A : Union[str, Any] ): '''simple docstring''' super().__init__(**A ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) UpperCAmelCase__ : int = kwargs.pop("""encoder""" ) UpperCAmelCase__ : int = encoder_config.pop("""model_type""" ) UpperCAmelCase__ : str = kwargs.pop("""decoder""" ) UpperCAmelCase__ : Dict = decoder_config.pop("""model_type""" ) UpperCAmelCase__ : List[Any] = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Any = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Union[str, Any] = True @classmethod def __lowercase ( cls : List[Any] ,A : PretrainedConfig ,A : PretrainedConfig ,**A : Tuple ): '''simple docstring''' logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : List[Any] = True return cls(encoder=encoder_config.to_dict() ,decoder=decoder_config.to_dict() ,**A ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Dict = self.encoder.to_dict() UpperCAmelCase__ : Any = self.decoder.to_dict() UpperCAmelCase__ : Dict = self.__class__.model_type return output class __lowercase ( __lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' return 1e-4 @property def __lowercase ( self : List[Any] ): '''simple docstring''' return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : Optional[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : List[str] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def __lowercase ( self : Dict ,A : "PreTrainedTokenizerBase" ,A : int = -1 ,A : int = -1 ,A : bool = False ,A : Optional["TensorType"] = None ,): '''simple docstring''' import torch UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : List[Any] = super().generate_dummy_inputs( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) UpperCAmelCase__ , UpperCAmelCase__ : int = dummy_input["""input_ids"""].shape UpperCAmelCase__ : int = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCAmelCase__ : Tuple = dummy_input.pop("""input_ids""" ) UpperCAmelCase__ : Optional[int] = dummy_input.pop("""attention_mask""" ) UpperCAmelCase__ : Dict = torch.zeros(A ) return common_inputs class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : Any ,A : PretrainedConfig ): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(A ) def __lowercase ( self : Dict ,A : PretrainedConfig ,A : PretrainedConfig ,A : str = "default" ): '''simple docstring''' UpperCAmelCase__ : List[str] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(A ,A )
65
1
"""simple docstring""" import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel __UpperCAmelCase = logging.getLogger(__name__) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if os.path.exists(__UpperCamelCase ): if os.path.exists(os.path.join(__UpperCamelCase , """config.json""" ) ) and os.path.isfile( os.path.join(__UpperCamelCase , """config.json""" ) ): os.remove(os.path.join(__UpperCamelCase , """config.json""" ) ) if os.path.exists(os.path.join(__UpperCamelCase , """pytorch_model.bin""" ) ) and os.path.isfile( os.path.join(__UpperCamelCase , """pytorch_model.bin""" ) ): os.remove(os.path.join(__UpperCamelCase , """pytorch_model.bin""" ) ) else: os.makedirs(__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase=False ): '''simple docstring''' UpperCAmelCase__ : str = 2 if unlogit: UpperCAmelCase__ : Tuple = torch.pow(__UpperCamelCase , __UpperCamelCase ) UpperCAmelCase__ : List[Any] = p * torch.log(__UpperCamelCase ) UpperCAmelCase__ : int = 0 return -plogp.sum(dim=-1 ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' logger.info("""lv, h >\t""" + """\t""".join(F"{x + 1}" for x in range(len(__UpperCamelCase ) ) ) ) for row in range(len(__UpperCamelCase ) ): if tensor.dtype != torch.long: logger.info(F"layer {row + 1}:\t" + """\t""".join(F"{x:.5f}" for x in tensor[row].cpu().data ) ) else: logger.info(F"layer {row + 1}:\t" + """\t""".join(F"{x:d}" for x in tensor[row].cpu().data ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=False ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = model.config.num_hidden_layers, model.config.num_attention_heads UpperCAmelCase__ : int = torch.zeros(__UpperCamelCase , __UpperCamelCase ).to(args.device ) UpperCAmelCase__ : Tuple = torch.zeros(__UpperCamelCase , __UpperCamelCase ).to(args.device ) if head_mask is None: UpperCAmelCase__ : Tuple = torch.ones(__UpperCamelCase , __UpperCamelCase ).to(args.device ) head_mask.requires_grad_(requires_grad=__UpperCamelCase ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: UpperCAmelCase__ : List[Any] = None UpperCAmelCase__ : Any = 0.0 UpperCAmelCase__ : List[str] = 0.0 for step, inputs in enumerate(tqdm(__UpperCamelCase , desc="""Iteration""" , disable=args.local_rank not in [-1, 0] ) ): UpperCAmelCase__ : Optional[int] = tuple(t.to(args.device ) for t in inputs ) ((UpperCAmelCase__) , ) : List[Any] = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) UpperCAmelCase__ : Tuple = model(__UpperCamelCase , labels=__UpperCamelCase , head_mask=__UpperCamelCase ) # (loss), lm_logits, presents, (all hidden_states), (attentions) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(__UpperCamelCase ): UpperCAmelCase__ : str = entropy(attn.detach() , __UpperCamelCase ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(__UpperCamelCase ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: UpperCAmelCase__ : Optional[Any] = 2 UpperCAmelCase__ : Union[str, Any] = torch.pow(torch.pow(__UpperCamelCase , __UpperCamelCase ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-20 if not args.dont_normalize_global_importance: UpperCAmelCase__ : Union[str, Any] = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info("""Attention entropies""" ) print_ad_tensor(__UpperCamelCase ) if compute_importance: logger.info("""Head importance scores""" ) print_ad_tensor(__UpperCamelCase ) logger.info("""Head ranked by importance scores""" ) UpperCAmelCase__ : Optional[Any] = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) UpperCAmelCase__ : Tuple = torch.arange( head_importance.numel() , device=args.device ) UpperCAmelCase__ : Optional[int] = head_ranks.view_as(__UpperCamelCase ) print_ad_tensor(__UpperCamelCase ) return attn_entropy, head_importance, total_loss def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : str = compute_heads_importance(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , compute_entropy=__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 1 / loss # instead of downsteam score use the LM loss logger.info("""Pruning: original score: %f, threshold: %f""" , __UpperCamelCase , original_score * args.masking_threshold ) UpperCAmelCase__ : List[Any] = torch.ones_like(__UpperCamelCase ) UpperCAmelCase__ : Any = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) UpperCAmelCase__ : Any = original_score while current_score >= original_score * args.masking_threshold: UpperCAmelCase__ : List[Any] = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads UpperCAmelCase__ : Optional[int] = float("""Inf""" ) UpperCAmelCase__ : List[Any] = head_importance.view(-1 ).sort()[1] if len(__UpperCamelCase ) <= num_to_mask: print("""BREAK BY num_to_mask""" ) break # mask heads UpperCAmelCase__ : List[str] = current_heads_to_mask[:num_to_mask] logger.info("""Heads to mask: %s""" , str(current_heads_to_mask.tolist() ) ) UpperCAmelCase__ : int = new_head_mask.view(-1 ) UpperCAmelCase__ : Dict = 0.0 UpperCAmelCase__ : Optional[int] = new_head_mask.view_as(__UpperCamelCase ) UpperCAmelCase__ : Optional[Any] = new_head_mask.clone().detach() print_ad_tensor(__UpperCamelCase ) # Compute metric and head importance again UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Dict = compute_heads_importance( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , compute_entropy=__UpperCamelCase , head_mask=__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 1 / loss logger.info( """Masking: current score: %f, remaining heads %d (%.1f percents)""" , __UpperCamelCase , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info("""Final head mask""" ) print_ad_tensor(__UpperCamelCase ) np.save(os.path.join(args.output_dir , """head_mask.npy""" ) , head_mask.detach().cpu().numpy() ) return head_mask def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = datetime.now() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Dict = compute_heads_importance( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , compute_entropy=__UpperCamelCase , compute_importance=__UpperCamelCase , head_mask=__UpperCamelCase ) UpperCAmelCase__ : int = 1 / loss UpperCAmelCase__ : int = datetime.now() - before_time UpperCAmelCase__ : int = sum(p.numel() for p in model.parameters() ) UpperCAmelCase__ : Optional[int] = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(__UpperCamelCase ) ) } for k, v in heads_to_prune.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : List[str] = [ v, ] assert sum(len(__UpperCamelCase ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(__UpperCamelCase ) UpperCAmelCase__ : List[Any] = sum(p.numel() for p in model.parameters() ) UpperCAmelCase__ : str = datetime.now() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = compute_heads_importance( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , compute_entropy=__UpperCamelCase , compute_importance=__UpperCamelCase , head_mask=__UpperCamelCase , actually_pruned=__UpperCamelCase , ) UpperCAmelCase__ : Optional[int] = 1 / loss UpperCAmelCase__ : List[str] = datetime.now() - before_time logger.info( """Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)""" , __UpperCamelCase , __UpperCamelCase , pruned_num_params / original_num_params * 100 , ) logger.info("""Pruning: score with masking: %f score with pruning: %f""" , __UpperCamelCase , __UpperCamelCase ) logger.info("""Pruning: speed ratio (original timing / new timing): %f percents""" , original_time / new_time * 100 ) save_model(__UpperCamelCase , args.output_dir ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--data_dir""" , default=__UpperCamelCase , type=__UpperCamelCase , required=__UpperCamelCase , help="""The input data dir. Should contain the .tsv files (or other data files) for the task.""" , ) parser.add_argument( """--model_name_or_path""" , default=__UpperCamelCase , type=__UpperCamelCase , required=__UpperCamelCase , help="""Path to pretrained model or model identifier from huggingface.co/models""" , ) parser.add_argument( """--output_dir""" , default=__UpperCamelCase , type=__UpperCamelCase , required=__UpperCamelCase , help="""The output directory where the model predictions and checkpoints will be written.""" , ) # Other parameters parser.add_argument( """--config_name""" , default="""""" , type=__UpperCamelCase , help="""Pretrained config name or path if not the same as model_name_or_path""" , ) parser.add_argument( """--tokenizer_name""" , default="""""" , type=__UpperCamelCase , help="""Pretrained tokenizer name or path if not the same as model_name_or_path""" , ) parser.add_argument( """--cache_dir""" , default=__UpperCamelCase , type=__UpperCamelCase , help="""Where do you want to store the pre-trained models downloaded from s3""" , ) parser.add_argument( """--data_subset""" , type=__UpperCamelCase , default=-1 , help="""If > 0: limit the data to a subset of data_subset instances.""" ) parser.add_argument( """--overwrite_output_dir""" , action="""store_true""" , help="""Whether to overwrite data in output directory""" ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) parser.add_argument( """--dont_normalize_importance_by_layer""" , action="""store_true""" , help="""Don't normalize importance score by layers""" ) parser.add_argument( """--dont_normalize_global_importance""" , action="""store_true""" , help="""Don't normalize all importance scores between 0 and 1""" , ) parser.add_argument( """--try_masking""" , action="""store_true""" , help="""Whether to try to mask head until a threshold of accuracy.""" ) parser.add_argument( """--masking_threshold""" , default=0.9 , type=__UpperCamelCase , help="""masking threshold in term of metrics (stop masking when metric < threshold * original metric value).""" , ) parser.add_argument( """--masking_amount""" , default=0.1 , type=__UpperCamelCase , help="""Amount to heads to masking at each masking step.""" ) parser.add_argument("""--metric_name""" , default="""acc""" , type=__UpperCamelCase , help="""Metric to use for head masking.""" ) parser.add_argument( """--max_seq_length""" , default=128 , type=__UpperCamelCase , help=( """The maximum total input sequence length after WordPiece tokenization. \n""" """Sequences longer than this will be truncated, sequences shorter padded.""" ) , ) parser.add_argument("""--batch_size""" , default=1 , type=__UpperCamelCase , help="""Batch size.""" ) parser.add_argument("""--seed""" , type=__UpperCamelCase , default=42 ) parser.add_argument("""--local_rank""" , type=__UpperCamelCase , default=-1 , help="""local_rank for distributed training on gpus""" ) parser.add_argument("""--no_cuda""" , action="""store_true""" , help="""Whether not to use CUDA when available""" ) parser.add_argument("""--server_ip""" , type=__UpperCamelCase , default="""""" , help="""Can be used for distant debugging.""" ) parser.add_argument("""--server_port""" , type=__UpperCamelCase , default="""""" , help="""Can be used for distant debugging.""" ) UpperCAmelCase__ : Any = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("""Waiting for debugger attach""" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=__UpperCamelCase ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: UpperCAmelCase__ : Dict = torch.device("""cuda""" if torch.cuda.is_available() and not args.no_cuda else """cpu""" ) UpperCAmelCase__ : List[str] = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) UpperCAmelCase__ : List[Any] = torch.device("""cuda""" , args.local_rank ) UpperCAmelCase__ : List[str] = 1 torch.distributed.init_process_group(backend="""nccl""" ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info("""device: {} n_gpu: {}, distributed: {}""".format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) UpperCAmelCase__ : Dict = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: UpperCAmelCase__ : List[Any] = nn.parallel.DistributedDataParallel( __UpperCamelCase , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=__UpperCamelCase ) elif args.n_gpu > 1: UpperCAmelCase__ : Dict = nn.DataParallel(__UpperCamelCase ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=__UpperCamelCase ) torch.save(__UpperCamelCase , os.path.join(args.output_dir , """run_args.bin""" ) ) logger.info("""Training/evaluation parameters %s""" , __UpperCamelCase ) # Prepare dataset UpperCAmelCase__ : str = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) UpperCAmelCase__ : Union[str, Any] = (torch.from_numpy(__UpperCamelCase ),) UpperCAmelCase__ : Optional[Any] = TensorDataset(*__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = RandomSampler(__UpperCamelCase ) UpperCAmelCase__ : Tuple = DataLoader(__UpperCamelCase , sampler=__UpperCamelCase , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: UpperCAmelCase__ : Any = mask_heads(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) prune_heads(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if __name__ == "__main__": main()
65
"""simple docstring""" import requests def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = {"""Content-Type""": """application/json"""} UpperCAmelCase__ : Optional[Any] = requests.post(__UpperCamelCase , json={"""text""": message_body} , headers=__UpperCamelCase ) if response.status_code != 200: UpperCAmelCase__ : Any = ( """Request to slack returned an error """ F"{response.status_code}, the response is:\n{response.text}" ) raise ValueError(__UpperCamelCase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
65
1
"""simple docstring""" import math class __lowercase : def __lowercase ( self : Any ,A : list[list[float]] ,A : list[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = 0.0 UpperCAmelCase__ : Any = 0.0 for i in range(len(A ) ): da += math.pow((sample[i] - weights[0][i]) ,2 ) da += math.pow((sample[i] - weights[1][i]) ,2 ) return 0 if da > da else 1 return 0 def __lowercase ( self : Union[str, Any] ,A : list[list[int | float]] ,A : list[int] ,A : int ,A : float ): '''simple docstring''' for i in range(len(A ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : int = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase__ : Union[str, Any] = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase__ : int = SelfOrganizingMap() UpperCAmelCase__ : Any = 3 UpperCAmelCase__ : Optional[int] = 0.5 for _ in range(__UpperCamelCase ): for j in range(len(__UpperCamelCase ) ): # training sample UpperCAmelCase__ : int = training_samples[j] # Compute the winning vector UpperCAmelCase__ : List[str] = self_organizing_map.get_winner(__UpperCamelCase , __UpperCamelCase ) # Update the winning vector UpperCAmelCase__ : Optional[Any] = self_organizing_map.update(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # classify test sample UpperCAmelCase__ : List[str] = [0, 0, 0, 1] UpperCAmelCase__ : Optional[int] = self_organizing_map.get_winner(__UpperCamelCase , __UpperCamelCase ) # results print(F"Clusters that the test sample belongs to : {winner}" ) print(F"Weights that have been trained : {weights}" ) # running the main() function if __name__ == "__main__": main()
65
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = CTRLTokenizer snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ : List[Any] = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase__ : Optional[int] = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : List[Any] = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase__ : int = {"""unk_token""": """<unk>"""} UpperCAmelCase__ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(A ) ) def __lowercase ( self : int ,**A : Dict ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """adapt react readapt apt""" UpperCAmelCase__ : Any = """adapt react readapt apt""" return input_text, output_text def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = CTRLTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) UpperCAmelCase__ : Tuple = """adapt react readapt apt""" UpperCAmelCase__ : Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase__ : Dict = tokenizer.tokenize(A ) self.assertListEqual(A ,A ) UpperCAmelCase__ : Any = tokens + [tokenizer.unk_token] UpperCAmelCase__ : Dict = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,A )
65
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['NllbTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['NllbTokenizerFast'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCAmelCase = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfig', 'BridgeTowerTextConfig', 'BridgeTowerVisionConfig', ], 'processing_bridgetower': ['BridgeTowerProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = ['BridgeTowerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST', 'BridgeTowerForContrastiveLearning', 'BridgeTowerForImageAndTextRetrieval', 'BridgeTowerForMaskedLM', 'BridgeTowerModel', 'BridgeTowerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
65
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaControlnetImgaImgPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """image""", """hint"""] snake_case_ = ["""image_embeds""", """negative_image_embeds""", """image""", """hint"""] snake_case_ = [ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __lowercase ( self : List[str] ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self : str ): '''simple docstring''' return 100 @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Optional[Any] = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } UpperCAmelCase__ : Optional[Any] = UNetaDConditionModel(**A ) return model @property def __lowercase ( self : List[str] ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : List[str] = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : str = self.dummy_unet UpperCAmelCase__ : Optional[Any] = self.dummy_movq UpperCAmelCase__ : Dict = { """num_train_timesteps""": 1_000, """beta_schedule""": """linear""", """beta_start""": 0.0_0_0_8_5, """beta_end""": 0.0_1_2, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } UpperCAmelCase__ : Any = DDIMScheduler(**A ) UpperCAmelCase__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __lowercase ( self : str ,A : Dict ,A : str=0 ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( A ) # create init_image UpperCAmelCase__ : Optional[int] = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Optional[int] = image.cpu().permute(0 ,2 ,3 ,1 )[0] UpperCAmelCase__ : Tuple = Image.fromarray(np.uinta(A ) ).convert("""RGB""" ).resize((256, 256) ) # create hint UpperCAmelCase__ : str = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) if str(A ).startswith("""mps""" ): UpperCAmelCase__ : List[Any] = torch.manual_seed(A ) else: UpperCAmelCase__ : Dict = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase__ : str = { """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : int = """cpu""" UpperCAmelCase__ : Optional[Any] = self.get_dummy_components() UpperCAmelCase__ : Dict = self.pipeline_class(**A ) UpperCAmelCase__ : Tuple = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[Any] = pipe(**self.get_dummy_inputs(A ) ) UpperCAmelCase__ : Union[str, Any] = output.images UpperCAmelCase__ : Optional[Any] = pipe( **self.get_dummy_inputs(A ) ,return_dict=A ,)[0] UpperCAmelCase__ : List[str] = image[0, -3:, -3:, -1] UpperCAmelCase__ : List[str] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Dict = np.array( [0.5_4_9_8_5_0_3_4, 0.5_5_5_0_9_3_6_5, 0.5_2_5_6_1_5_0_4, 0.5_5_7_0_4_9_4, 0.5_5_9_3_8_1_8, 0.5_2_6_3_9_7_9, 0.5_0_2_8_5_6_4_3, 0.5_0_6_9_8_4_6, 0.5_1_1_9_6_7_3_6] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : Any ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy""" ) UpperCAmelCase__ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) UpperCAmelCase__ : Union[str, Any] = init_image.resize((512, 512) ) UpperCAmelCase__ : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) UpperCAmelCase__ : List[str] = torch.from_numpy(np.array(A ) ).float() / 2_5_5.0 UpperCAmelCase__ : Tuple = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) UpperCAmelCase__ : Union[str, Any] = """A robot, 4k photo""" UpperCAmelCase__ : Dict = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" ,torch_dtype=torch.floataa ) pipe_prior.to(A ) UpperCAmelCase__ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" ,torch_dtype=torch.floataa ) UpperCAmelCase__ : Dict = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = pipe_prior( A ,image=A ,strength=0.8_5 ,generator=A ,negative_prompt="""""" ,).to_tuple() UpperCAmelCase__ : Optional[int] = pipeline( image=A ,image_embeds=A ,negative_image_embeds=A ,hint=A ,generator=A ,num_inference_steps=100 ,height=512 ,width=512 ,strength=0.5 ,output_type="""np""" ,) UpperCAmelCase__ : int = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A ,A )
65
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """is_longer"""] def __init__( self : str ,A : Union[str, Any]=64 ,A : Tuple=48_000 ,A : Dict=480 ,A : List[str]=10 ,A : str=1_024 ,A : Any=0.0 ,A : Optional[int]=False ,A : float = 0 ,A : float = 14_000 ,A : int = None ,A : str = "fusion" ,A : str = "repeatpad" ,**A : List[Any] ,): '''simple docstring''' super().__init__( feature_size=A ,sampling_rate=A ,padding_value=A ,return_attention_mask=A ,**A ,) UpperCAmelCase__ : List[Any] = top_db UpperCAmelCase__ : Union[str, Any] = truncation UpperCAmelCase__ : Optional[int] = padding UpperCAmelCase__ : List[Any] = fft_window_size UpperCAmelCase__ : Optional[Any] = (fft_window_size >> 1) + 1 UpperCAmelCase__ : Any = hop_length UpperCAmelCase__ : List[str] = max_length_s UpperCAmelCase__ : List[Any] = max_length_s * sampling_rate UpperCAmelCase__ : List[Any] = sampling_rate UpperCAmelCase__ : Optional[int] = frequency_min UpperCAmelCase__ : Tuple = frequency_max UpperCAmelCase__ : List[str] = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm=A ,mel_scale="""htk""" ,) UpperCAmelCase__ : str = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins ,num_mel_filters=A ,min_frequency=A ,max_frequency=A ,sampling_rate=A ,norm="""slaney""" ,mel_scale="""slaney""" ,) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def __lowercase ( self : List[str] ,A : np.array ,A : Optional[np.array] = None ): '''simple docstring''' UpperCAmelCase__ : Dict = spectrogram( A ,window_function(self.fft_window_size ,"""hann""" ) ,frame_length=self.fft_window_size ,hop_length=self.hop_length ,power=2.0 ,mel_filters=A ,log_mel="""dB""" ,) return log_mel_spectrogram.T def __lowercase ( self : Optional[Any] ,A : Union[str, Any] ,A : int ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = np.array_split(list(range(0 ,total_frames - chunk_frames + 1 ) ) ,3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : List[str] = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase__ : int = [0] # randomly choose index for each part UpperCAmelCase__ : Tuple = np.random.choice(ranges[0] ) UpperCAmelCase__ : Tuple = np.random.choice(ranges[1] ) UpperCAmelCase__ : str = np.random.choice(ranges[2] ) UpperCAmelCase__ : List[str] = mel[idx_front : idx_front + chunk_frames, :] UpperCAmelCase__ : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] UpperCAmelCase__ : Dict = mel[idx_back : idx_back + chunk_frames, :] UpperCAmelCase__ : Optional[Any] = torch.tensor(mel[None, None, :] ) UpperCAmelCase__ : int = torch.nn.functional.interpolate( A ,size=[chunk_frames, 64] ,mode="""bilinear""" ,align_corners=A ) UpperCAmelCase__ : Dict = mel_shrink[0][0].numpy() UpperCAmelCase__ : Dict = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] ,axis=0 ) return mel_fusion def __lowercase ( self : Any ,A : np.array ,A : Optional[int] ,A : Any ,A : Tuple ): '''simple docstring''' if waveform.shape[0] > max_length: if truncation == "rand_trunc": UpperCAmelCase__ : int = True # random crop to max_length (for compatibility) -> this should be handled by self.pad UpperCAmelCase__ : str = len(A ) - max_length UpperCAmelCase__ : Optional[Any] = np.random.randint(0 ,overflow + 1 ) UpperCAmelCase__ : Optional[int] = waveform[idx : idx + max_length] UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] elif truncation == "fusion": UpperCAmelCase__ : Tuple = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : Optional[int] = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed UpperCAmelCase__ : int = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. UpperCAmelCase__ : List[Any] = np.stack([mel, mel, mel, mel] ,axis=0 ) UpperCAmelCase__ : Any = False else: UpperCAmelCase__ : Union[str, Any] = self._random_mel_fusion(A ,A ,A ) UpperCAmelCase__ : List[str] = True else: raise NotImplementedError(f"data_truncating {truncation} not implemented" ) else: UpperCAmelCase__ : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": UpperCAmelCase__ : str = int(max_length / len(A ) ) UpperCAmelCase__ : int = np.stack(np.tile(A ,n_repeat + 1 ) )[:max_length] if padding == "repeatpad": UpperCAmelCase__ : List[Any] = int(max_length / len(A ) ) UpperCAmelCase__ : str = np.stack(np.tile(A ,A ) ) UpperCAmelCase__ : Optional[Any] = np.pad(A ,(0, max_length - waveform.shape[0]) ,mode="""constant""" ,constant_values=0 ) if truncation == "fusion": UpperCAmelCase__ : int = self._np_extract_fbank_features(A ,self.mel_filters ) UpperCAmelCase__ : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] ,axis=0 ) else: UpperCAmelCase__ : Any = self._np_extract_fbank_features(A ,self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self : str ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : str = None ,A : Optional[str] = None ,A : Optional[int] = None ,A : Optional[int] = None ,A : Optional[Union[str, TensorType]] = None ,**A : List[str] ,): '''simple docstring''' UpperCAmelCase__ : Optional[int] = truncation if truncation is not None else self.truncation UpperCAmelCase__ : Dict = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" f" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" f" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[int] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : List[str] = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Any = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [np.asarray(A )] # convert to mel spectrogram, truncate and pad if needed. UpperCAmelCase__ : Tuple = [ self._get_input_mel(A ,max_length if max_length else self.nb_max_samples ,A ,A ) for waveform in raw_speech ] UpperCAmelCase__ : Optional[int] = [] UpperCAmelCase__ : Tuple = [] for mel, longer in padded_inputs: input_mel.append(A ) is_longer.append(A ) if truncation == "fusion" and sum(A ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer UpperCAmelCase__ : List[str] = np.random.randint(0 ,len(A ) ) UpperCAmelCase__ : int = True if isinstance(input_mel[0] ,A ): UpperCAmelCase__ : Tuple = [np.asarray(A ,dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool UpperCAmelCase__ : List[str] = [[longer] for longer in is_longer] UpperCAmelCase__ : List[Any] = {"""input_features""": input_mel, """is_longer""": is_longer} UpperCAmelCase__ : str = BatchFeature(A ) if return_tensors is not None: UpperCAmelCase__ : int = input_features.convert_to_tensors(A ) return input_features
65
1
"""simple docstring""" from __future__ import annotations from cmath import sqrt def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if a == 0: raise ValueError("""Coefficient 'a' must not be zero.""" ) UpperCAmelCase__ : List[Any] = b * b - 4 * a * c UpperCAmelCase__ : Any = (-b + sqrt(__UpperCamelCase )) / (2 * a) UpperCAmelCase__ : Optional[Any] = (-b - sqrt(__UpperCamelCase )) / (2 * a) return ( root_a.real if not root_a.imag else root_a, root_a.real if not root_a.imag else root_a, ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = quadratic_roots(a=5 , b=6 , c=1 ) print(F"The solutions are: {solutiona} and {solutiona}" ) if __name__ == "__main__": main()
65
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Union[str, Any] ,A : Dict=7 ,A : Optional[int]=3 ,A : List[str]=18 ,A : Union[str, Any]=30 ,A : Tuple=400 ,A : Dict=True ,A : List[str]=None ,A : str=True ,A : Optional[Any]=False ,A : Optional[Any]=True ,A : List[str]=True ,A : Optional[int]=[0.5, 0.5, 0.5] ,A : List[str]=[0.5, 0.5, 0.5] ,): '''simple docstring''' UpperCAmelCase__ : str = parent UpperCAmelCase__ : List[str] = batch_size UpperCAmelCase__ : List[str] = num_channels UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : Optional[int] = max_resolution UpperCAmelCase__ : str = do_resize UpperCAmelCase__ : Tuple = size if size is not None else {"""height""": 18, """width""": 20} UpperCAmelCase__ : List[str] = do_thumbnail UpperCAmelCase__ : Optional[int] = do_align_axis UpperCAmelCase__ : Union[str, Any] = do_pad UpperCAmelCase__ : Tuple = do_normalize UpperCAmelCase__ : Optional[Any] = image_mean UpperCAmelCase__ : List[Any] = image_std def __lowercase ( self : Optional[int] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = DonutImageProcessor if is_vision_available() else None def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = DonutImageProcessingTester(self ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_thumbnail""" ) ) self.assertTrue(hasattr(A ,"""do_align_long_axis""" ) ) self.assertTrue(hasattr(A ,"""do_pad""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 18, """width""": 20} ) UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order UpperCAmelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ,size=(42, 84) ) self.assertEqual(image_processor.size ,{"""height""": 84, """width""": 42} ) def __lowercase ( self : Dict ): '''simple docstring''' pass @is_flaky() def __lowercase ( self : int ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase__ : int = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Tuple = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : List[str] ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase__ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : Optional[int] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) @is_flaky() def __lowercase ( self : Any ): '''simple docstring''' # Initialize image_processing UpperCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase__ : List[Any] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched UpperCAmelCase__ : List[Any] = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,)
65
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=__lowerCamelCase ): snake_case_ = ["""onnx"""] def __init__( self : int ,*A : List[str] ,**A : int ): '''simple docstring''' requires_backends(self ,["""onnx"""] ) @classmethod def __lowercase ( cls : Optional[Any] ,*A : List[str] ,**A : Dict ): '''simple docstring''' requires_backends(cls ,["""onnx"""] ) @classmethod def __lowercase ( cls : List[Any] ,*A : Optional[int] ,**A : int ): '''simple docstring''' requires_backends(cls ,["""onnx"""] )
65
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class __lowercase ( __lowerCamelCase ): snake_case_ = """open-llama""" def __init__( self : Dict ,A : str=100_000 ,A : str=4_096 ,A : Optional[Any]=11_008 ,A : Tuple=32 ,A : str=32 ,A : Optional[int]="silu" ,A : List[Any]=2_048 ,A : str=0.0_2 ,A : Optional[int]=1e-6 ,A : int=True ,A : Tuple=0 ,A : str=1 ,A : Any=2 ,A : Optional[Any]=False ,A : int=True ,A : Any=0.1 ,A : Optional[Any]=0.1 ,A : Optional[Any]=True ,A : Union[str, Any]=True ,A : Tuple=None ,**A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : str = vocab_size UpperCAmelCase__ : List[str] = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : Tuple = intermediate_size UpperCAmelCase__ : Optional[int] = num_hidden_layers UpperCAmelCase__ : Any = num_attention_heads UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = initializer_range UpperCAmelCase__ : Optional[int] = rms_norm_eps UpperCAmelCase__ : Any = use_cache UpperCAmelCase__ : Optional[Any] = kwargs.pop( """use_memorry_efficient_attention""" ,A ) UpperCAmelCase__ : Any = hidden_dropout_prob UpperCAmelCase__ : str = attention_dropout_prob UpperCAmelCase__ : Optional[int] = use_stable_embedding UpperCAmelCase__ : Tuple = shared_input_output_embedding UpperCAmelCase__ : Tuple = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,tie_word_embeddings=A ,**A ,) def __lowercase ( self : Optional[Any] ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ f"got {self.rope_scaling}" ) UpperCAmelCase__ : List[Any] = self.rope_scaling.get("""type""" ,A ) UpperCAmelCase__ : int = self.rope_scaling.get("""factor""" ,A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(A ,A ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
65
1
"""simple docstring""" from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
65
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments __UpperCAmelCase = logging.getLogger(__name__) @dataclass class __lowercase ( __lowerCamelCase ): snake_case_ = field( default=0.0 , metadata={"""help""": """The label smoothing epsilon to apply (if not zero)."""} ) snake_case_ = field(default=__lowerCamelCase , metadata={"""help""": """Whether to SortishSamler or not."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Whether to use generate to calculate generative metrics (ROUGE, BLEU)."""} ) snake_case_ = field(default=__lowerCamelCase , metadata={"""help""": """whether to use adafactor"""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Encoder layer dropout probability. Goes into model.config."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Decoder layer dropout probability. Goes into model.config."""} ) snake_case_ = field(default=__lowerCamelCase , metadata={"""help""": """Dropout probability. Goes into model.config."""} ) snake_case_ = field( default=__lowerCamelCase , metadata={"""help""": """Attention dropout probability. Goes into model.config."""} ) snake_case_ = field( default="""linear""" , metadata={"""help""": F"Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}"} , )
65
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = ["""input_features""", """attention_mask"""] def __init__( self : Any ,A : str=80 ,A : Optional[int]=16_000 ,A : int=0.0 ,A : str=10 ,A : Any=25 ,A : str="hamming_window" ,A : int=3_2_7_6_8.0 ,A : List[str]=0.9_7 ,A : Optional[int]=1.0 ,A : Optional[Any]=True ,A : Tuple=True ,A : Any=False ,**A : int ,): '''simple docstring''' super().__init__(feature_size=A ,sampling_rate=A ,padding_value=A ,**A ) UpperCAmelCase__ : str = feature_size UpperCAmelCase__ : int = sampling_rate UpperCAmelCase__ : int = padding_value UpperCAmelCase__ : Dict = hop_length UpperCAmelCase__ : int = win_length UpperCAmelCase__ : Dict = frame_signal_scale UpperCAmelCase__ : Dict = preemphasis_coeff UpperCAmelCase__ : str = mel_floor UpperCAmelCase__ : Any = normalize_means UpperCAmelCase__ : str = normalize_vars UpperCAmelCase__ : int = win_function UpperCAmelCase__ : List[Any] = return_attention_mask UpperCAmelCase__ : str = win_length * sampling_rate // 1_000 UpperCAmelCase__ : List[Any] = hop_length * sampling_rate // 1_000 UpperCAmelCase__ : int = optimal_fft_length(self.sample_size ) UpperCAmelCase__ : List[Any] = (self.n_fft // 2) + 1 def __lowercase ( self : Union[str, Any] ,A : np.array ): '''simple docstring''' if self.win_function == "hamming_window": UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=A ) else: UpperCAmelCase__ : Any = window_function(window_length=self.sample_size ,name=self.win_function ) UpperCAmelCase__ : Union[str, Any] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) UpperCAmelCase__ : Optional[Any] = spectrogram( one_waveform * self.frame_signal_scale ,window=A ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=A ,preemphasis=self.preemphasis_coeff ,mel_filters=A ,mel_floor=self.mel_floor ,log_mel="""log""" ,) return msfc_features.T def __lowercase ( self : str ,A : Any ,A : Optional[int] ,A : str ): '''simple docstring''' # make sure we normalize float32 arrays if self.normalize_means: UpperCAmelCase__ : Optional[Any] = x[:input_length].mean(axis=0 ) UpperCAmelCase__ : Any = np.subtract(A ,A ) if self.normalize_vars: UpperCAmelCase__ : str = x[:input_length].std(axis=0 ) UpperCAmelCase__ : Optional[int] = np.divide(A ,A ) if input_length < x.shape[0]: UpperCAmelCase__ : int = padding_value # make sure array is in float32 UpperCAmelCase__ : str = x.astype(np.floataa ) return x def __lowercase ( self : Union[str, Any] ,A : List[np.ndarray] ,A : Optional[np.ndarray] = None ): '''simple docstring''' UpperCAmelCase__ : Any = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(A ,A ,self.padding_value ) for x, n in zip(A ,A )] def __call__( self : Union[str, Any] ,A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,A : Union[bool, str, PaddingStrategy] = False ,A : Optional[int] = None ,A : bool = False ,A : Optional[int] = None ,A : Optional[bool] = None ,A : Optional[Union[str, TensorType]] = None ,A : Optional[int] = None ,**A : Tuple ,): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" f" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" f" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase__ : Optional[Any] = isinstance(A ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCAmelCase__ : Any = is_batched_numpy or ( isinstance(A ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ : List[str] = [np.asarray(A ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A ,np.ndarray ): UpperCAmelCase__ : Union[str, Any] = np.asarray(A ,dtype=np.floataa ) elif isinstance(A ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ : Optional[Any] = [raw_speech] # extract fbank features UpperCAmelCase__ : Tuple = [self._extract_mfsc_features(A ) for one_waveform in raw_speech] # convert into correct format for padding UpperCAmelCase__ : str = BatchFeature({"""input_features""": features} ) UpperCAmelCase__ : Optional[Any] = self.pad( A ,padding=A ,max_length=A ,truncation=A ,pad_to_multiple_of=A ,return_attention_mask=A ,**A ,) # make sure list is in array format UpperCAmelCase__ : Tuple = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] ,A ): UpperCAmelCase__ : Union[str, Any] = [np.asarray(A ,dtype=np.floataa ) for feature in input_features] UpperCAmelCase__ : Dict = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: UpperCAmelCase__ : str = [np.asarray(A ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: UpperCAmelCase__ : Union[str, Any] = ( np.array(A ,dtype=np.intaa ) if self._get_padding_strategies(A ,max_length=A ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) UpperCAmelCase__ : Any = self.normalize( padded_inputs["""input_features"""] ,attention_mask=A ) if return_tensors is not None: UpperCAmelCase__ : Union[str, Any] = padded_inputs.convert_to_tensors(A ) return padded_inputs
65
1
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrajectoryTransformerModel', 'TrajectoryTransformerPreTrainedModel', 'load_tf_weights_in_trajectory_transformer', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
"""simple docstring""" from math import factorial def lowerCAmelCase ( __UpperCamelCase = 100 ): '''simple docstring''' return sum(int(__UpperCamelCase ) for x in str(factorial(__UpperCamelCase ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
65
1
"""simple docstring""" import re def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = re.compile(r"""^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$""" ) if match := re.search(__UpperCamelCase , __UpperCamelCase ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
65
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] ,A : Optional[int] ,A : int=13 ,A : Tuple=7 ,A : Dict=True ,A : Optional[int]=True ,A : Tuple=True ,A : str=True ,A : Any=99 ,A : Tuple=32 ,A : Dict=5 ,A : Optional[int]=4 ,A : Dict=37 ,A : Any="gelu" ,A : Any=0.1 ,A : Optional[int]=0.1 ,A : Union[str, Any]=512 ,A : Any=16 ,A : List[str]=2 ,A : List[Any]=0.0_2 ,A : Optional[int]=4 ,): '''simple docstring''' UpperCAmelCase__ : Dict = parent UpperCAmelCase__ : Any = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : Optional[int] = is_training UpperCAmelCase__ : Optional[Any] = use_attention_mask UpperCAmelCase__ : int = use_token_type_ids UpperCAmelCase__ : int = use_labels UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Any = hidden_act UpperCAmelCase__ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : List[Any] = type_vocab_size UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : List[Any] = num_choices def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : List[str] = None if self.use_attention_mask: UpperCAmelCase__ : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : int = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=A ,) return config, input_ids, attention_mask def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = config_and_inputs UpperCAmelCase__ : str = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = FlaxDistilBertModelTester(self ) @slow def __lowercase ( self : Optional[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = model_class_name.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(A ) @require_flax class __lowercase ( unittest.TestCase ): @slow def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) UpperCAmelCase__ : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) UpperCAmelCase__ : str = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ : Dict = model(A ,attention_mask=A )[0] UpperCAmelCase__ : List[Any] = (1, 11, 768) self.assertEqual(output.shape ,A ) UpperCAmelCase__ : Any = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,A ,atol=1e-4 ) )
65
1
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = MobileBertConfig.from_json_file(__UpperCamelCase ) print(F"Building PyTorch model from configuration: {config}" ) UpperCAmelCase__ : Union[str, Any] = MobileBertForPreTraining(__UpperCamelCase ) # Load weights from tf checkpoint UpperCAmelCase__ : Union[str, Any] = load_tf_weights_in_mobilebert(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , __UpperCamelCase ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--mobilebert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained MobileBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __UpperCAmelCase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
65
"""simple docstring""" __UpperCAmelCase = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image']) __UpperCAmelCase = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __UpperCAmelCase = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __UpperCAmelCase = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['image', 'mask_image']) __UpperCAmelCase = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __UpperCAmelCase = frozenset(['example_image', 'image', 'mask_image']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['class_labels']) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset(['batch_size']) __UpperCAmelCase = frozenset([]) __UpperCAmelCase = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __UpperCAmelCase = frozenset(['prompt', 'negative_prompt']) __UpperCAmelCase = frozenset(['input_tokens']) __UpperCAmelCase = frozenset(['input_tokens'])
65
1
"""simple docstring""" from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class __lowercase ( __lowerCamelCase ): def __init__( self : int ,A : Optional[NestedDataStructureLike[PathLike]] = None ,A : Optional[NamedSplit] = None ,A : Optional[Features] = None ,A : str = None ,A : bool = False ,A : bool = False ,A : Optional[int] = None ,**A : List[Any] ,): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = path_or_paths UpperCAmelCase__ : Dict = split if split or isinstance(A ,A ) else """train""" UpperCAmelCase__ : Dict = features UpperCAmelCase__ : Dict = cache_dir UpperCAmelCase__ : Tuple = keep_in_memory UpperCAmelCase__ : Optional[int] = streaming UpperCAmelCase__ : Any = num_proc UpperCAmelCase__ : List[Any] = kwargs @abstractmethod def __lowercase ( self : Any ): '''simple docstring''' pass class __lowercase ( __lowerCamelCase ): def __init__( self : Optional[int] ,A : Optional[Features] = None ,A : str = None ,A : bool = False ,A : bool = False ,A : Optional[int] = None ,**A : Optional[int] ,): '''simple docstring''' UpperCAmelCase__ : List[str] = features UpperCAmelCase__ : str = cache_dir UpperCAmelCase__ : Union[str, Any] = keep_in_memory UpperCAmelCase__ : Optional[Any] = streaming UpperCAmelCase__ : Optional[Any] = num_proc UpperCAmelCase__ : Dict = kwargs @abstractmethod def __lowercase ( self : List[Any] ): '''simple docstring''' pass
65
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Dict = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase__ : Tuple = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[Any] = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase__ : int = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 16_000, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase__ : Optional[int] = tempfile.mkdtemp() UpperCAmelCase__ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,A ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.feature_extraction_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) # load decoder from hub UpperCAmelCase__ : int = """hf-internal-testing/ngram-beam-search-decoder""" def __lowercase ( self : str ,**A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.add_kwargs_tokens_map.copy() kwargs.update(A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : List[str] ,**A : Dict ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Any ,**A : List[Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name ,**A ) def __lowercase ( self : Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : str = self.get_decoder() UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() ,feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor ,A ) # decoder self.assertEqual(processor.decoder._alphabet.labels ,decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set ,decoder.model_container[decoder._model_key]._unigram_set ,) self.assertIsInstance(processor.decoder ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase__ : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname ,alpha=5.0 ,beta=3.0 ,score_boundary=-7.0 ,unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha ,5.0 ) self.assertEqual(processor.language_model.beta ,3.0 ) self.assertEqual(processor.language_model.score_boundary ,-7.0 ) self.assertEqual(processor.language_model.unk_score_offset ,3 ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(A ,"""include""" ): WavaVecaProcessorWithLM( tokenizer=A ,feature_extractor=self.get_feature_extractor() ,decoder=self.get_decoder() ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Any = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : str = floats_list((3, 1_000) ) UpperCAmelCase__ : Optional[Any] = feature_extractor(A ,return_tensors="""np""" ) UpperCAmelCase__ : List[Any] = processor(A ,return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1e-2 ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = self.get_feature_extractor() UpperCAmelCase__ : Union[str, Any] = self.get_tokenizer() UpperCAmelCase__ : Optional[int] = self.get_decoder() UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : List[Any] = """This is a test string""" UpperCAmelCase__ : int = processor(text=A ) UpperCAmelCase__ : Dict = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def __lowercase ( self : Tuple ,A : List[Any]=(2, 10, 16) ,A : Dict=77 ): '''simple docstring''' np.random.seed(A ) return np.random.rand(*A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits(shape=(10, 16) ,seed=13 ) UpperCAmelCase__ : Tuple = processor.decode(A ) UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams(A )[0] self.assertEqual(decoded_decoder[0] ,decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" ,decoded_processor.text ) self.assertEqual(decoded_decoder[-2] ,decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] ,decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def __lowercase ( self : List[str] ,A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_feature_extractor() UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Optional[Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase__ : List[str] = processor.batch_decode(A ) else: with get_context(A ).Pool() as pool: UpperCAmelCase__ : Union[str, Any] = processor.batch_decode(A ,A ) UpperCAmelCase__ : Optional[Any] = list(A ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase__ : Union[str, Any] = decoder.decode_beams_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(A ,decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] ,decoded_processor.text ) self.assertListEqual(A ,decoded_processor.logit_score ) self.assertListEqual(A ,decoded_processor.lm_score ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Any = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Dict = self._get_dummy_logits() UpperCAmelCase__ : Any = 15 UpperCAmelCase__ : Dict = -2_0.0 UpperCAmelCase__ : List[Any] = -4.0 UpperCAmelCase__ : Union[str, Any] = processor.batch_decode( A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[str] = decoded_processor_out.text UpperCAmelCase__ : List[str] = list(A ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : Tuple = decoder.decode_beams_batch( A ,A ,beam_width=A ,beam_prune_logp=A ,token_min_logp=A ,) UpperCAmelCase__ : List[Any] = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase__ : Any = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase__ : List[str] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] ,A ) self.assertTrue(np.array_equal(A ,decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-2_0.0_5_4, -1_8.4_4_7] ,A ,atol=1e-3 ) ) self.assertTrue(np.array_equal(A ,decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-1_5.5_5_4, -1_3.9_4_7_4] ,A ,atol=1e-3 ) ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_feature_extractor() UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : int = self.get_decoder() UpperCAmelCase__ : str = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) UpperCAmelCase__ : Tuple = self._get_dummy_logits() UpperCAmelCase__ : Tuple = 2.0 UpperCAmelCase__ : str = 5.0 UpperCAmelCase__ : Union[str, Any] = -2_0.0 UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = processor.batch_decode( A ,alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) UpperCAmelCase__ : Any = decoded_processor_out.text UpperCAmelCase__ : Union[str, Any] = list(A ) decoder.reset_params( alpha=A ,beta=A ,unk_score_offset=A ,lm_score_boundary=A ,) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase__ : List[Any] = decoder.decode_beams_batch( A ,A ,) UpperCAmelCase__ : Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(A ,A ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] ,A ) UpperCAmelCase__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha ,2.0 ) self.assertEqual(lm_model.beta ,5.0 ) self.assertEqual(lm_model.unk_score_offset ,-2_0.0 ) self.assertEqual(lm_model.score_boundary ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : str = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Any = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Optional[int] = os.listdir(A ) UpperCAmelCase__ : List[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(A ,A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(A ) UpperCAmelCase__ : Tuple = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase__ : Optional[int] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase__ : Tuple = os.listdir(A ) UpperCAmelCase__ : Dict = os.listdir(A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(A ,A ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = floats_list((3, 1_000) ) UpperCAmelCase__ : List[str] = processor_wavaveca(A ,return_tensors="""np""" ) UpperCAmelCase__ : Dict = processor_auto(A ,return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() ,input_auto[key].sum() ,delta=1e-2 ) UpperCAmelCase__ : List[str] = self._get_dummy_logits() UpperCAmelCase__ : Tuple = processor_wavaveca.batch_decode(A ) UpperCAmelCase__ : List[str] = processor_auto.batch_decode(A ) self.assertListEqual(decoded_wavaveca.text ,decoded_auto.text ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.get_feature_extractor() UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : List[Any] = self.get_decoder() UpperCAmelCase__ : int = WavaVecaProcessorWithLM(tokenizer=A ,feature_extractor=A ,decoder=A ) self.assertListEqual( processor.model_input_names ,feature_extractor.model_input_names ,msg="""`processor` and `feature_extractor` model input names do not match""" ,) @staticmethod def __lowercase ( A : Optional[Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : Dict = self._get_dummy_logits()[0] UpperCAmelCase__ : List[str] = processor.decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ) ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] ,"""end_offset""" ) ,[1, 3, 5] ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase__ : int = self._get_dummy_logits() UpperCAmelCase__ : Any = processor.batch_decode(A ,output_word_offsets=A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) ,4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(A ,"""word""" ) ) for o in outputs["""word_offsets"""]] ,outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""word""" ) ,["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""start_offset""" ) ,[0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] ,"""end_offset""" ) ,[1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self : Tuple ): '''simple docstring''' import torch UpperCAmelCase__ : Any = load_dataset("""common_voice""" ,"""en""" ,split="""train""" ,streaming=A ) UpperCAmelCase__ : Tuple = ds.cast_column("""audio""" ,datasets.Audio(sampling_rate=16_000 ) ) UpperCAmelCase__ : Tuple = iter(A ) UpperCAmelCase__ : Optional[int] = next(A ) UpperCAmelCase__ : List[Any] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase__ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase__ : Tuple = processor(sample["""audio"""]["""array"""] ,return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(A ).logits.cpu().numpy() UpperCAmelCase__ : Any = processor.decode(logits[0] ,output_word_offsets=A ) UpperCAmelCase__ : str = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase__ : Dict = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,A ) self.assertEqual(""" """.join(self.get_from_offsets(A ,"""word""" ) ) ,output.text ) # output times UpperCAmelCase__ : str = torch.tensor(self.get_from_offsets(A ,"""start_time""" ) ) UpperCAmelCase__ : List[Any] = torch.tensor(self.get_from_offsets(A ,"""end_time""" ) ) # fmt: off UpperCAmelCase__ : Union[str, Any] = torch.tensor([1.4_1_9_9, 1.6_5_9_9, 2.2_5_9_9, 3.0, 3.2_4, 3.5_9_9_9, 3.7_9_9_9, 4.0_9_9_9, 4.2_6, 4.9_4, 5.2_8, 5.6_5_9_9, 5.7_8, 5.9_4, 6.3_2, 6.5_3_9_9, 6.6_5_9_9] ) UpperCAmelCase__ : List[Any] = torch.tensor([1.5_3_9_9, 1.8_9_9_9, 2.9, 3.1_6, 3.5_3_9_9, 3.7_2, 4.0_1_9_9, 4.1_7_9_9, 4.7_6, 5.1_5_9_9, 5.5_5_9_9, 5.6_9_9_9, 5.8_6, 6.1_9_9_9, 6.3_8, 6.6_1_9_9, 6.9_4] ) # fmt: on self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) ) self.assertTrue(torch.allclose(A ,A ,atol=0.0_1 ) )
65
1
"""simple docstring""" import os def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : str = os.path.join(os.path.dirname(__UpperCamelCase ) , """num.txt""" ) with open(__UpperCamelCase ) as file_hand: return str(sum(int(__UpperCamelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
65
"""simple docstring""" from sklearn.metrics import fa_score import datasets __UpperCAmelCase = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __UpperCAmelCase = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __UpperCAmelCase = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def __lowercase ( self : List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) ,reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"""] ,) def __lowercase ( self : Union[str, Any] ,A : List[str] ,A : List[Any] ,A : Optional[Any]=None ,A : List[str]=1 ,A : Optional[Any]="binary" ,A : Any=None ): '''simple docstring''' UpperCAmelCase__ : List[Any] = fa_score( A ,A ,labels=A ,pos_label=A ,average=A ,sample_weight=A ) return {"f1": float(A ) if score.size == 1 else score}
65
1
"""simple docstring""" from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class __lowercase ( __lowerCamelCase ): snake_case_ = 42 snake_case_ = 42 class __lowercase ( nn.Module ): snake_case_ = 42 snake_case_ = (1_6, 3_2, 9_6, 2_5_6) snake_case_ = jnp.floataa def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = nn.Conv( self.block_out_channels[0] ,kernel_size=(3, 3) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) UpperCAmelCase__ : Tuple = [] for i in range(len(self.block_out_channels ) - 1 ): UpperCAmelCase__ : Union[str, Any] = self.block_out_channels[i] UpperCAmelCase__ : Dict = self.block_out_channels[i + 1] UpperCAmelCase__ : Optional[int] = nn.Conv( A ,kernel_size=(3, 3) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) blocks.append(A ) UpperCAmelCase__ : Optional[int] = nn.Conv( A ,kernel_size=(3, 3) ,strides=(2, 2) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) blocks.append(A ) UpperCAmelCase__ : List[Any] = blocks UpperCAmelCase__ : Union[str, Any] = nn.Conv( self.conditioning_embedding_channels ,kernel_size=(3, 3) ,padding=((1, 1), (1, 1)) ,kernel_init=nn.initializers.zeros_init() ,bias_init=nn.initializers.zeros_init() ,dtype=self.dtype ,) def __call__( self : Dict ,A : int ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.conv_in(A ) UpperCAmelCase__ : Tuple = nn.silu(A ) for block in self.blocks: UpperCAmelCase__ : Tuple = block(A ) UpperCAmelCase__ : str = nn.silu(A ) UpperCAmelCase__ : Dict = self.conv_out(A ) return embedding @flax_register_to_config class __lowercase ( nn.Module , __lowerCamelCase , __lowerCamelCase ): snake_case_ = 3_2 snake_case_ = 4 snake_case_ = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) snake_case_ = False snake_case_ = (3_2_0, 6_4_0, 1_2_8_0, 1_2_8_0) snake_case_ = 2 snake_case_ = 8 snake_case_ = None snake_case_ = 1_2_8_0 snake_case_ = 0.0 snake_case_ = False snake_case_ = jnp.floataa snake_case_ = True snake_case_ = 0 snake_case_ = "rgb" snake_case_ = (1_6, 3_2, 9_6, 2_5_6) def __lowercase ( self : Tuple ,A : jax.random.KeyArray ): '''simple docstring''' # init input tensors UpperCAmelCase__ : Optional[int] = (1, self.in_channels, self.sample_size, self.sample_size) UpperCAmelCase__ : Optional[Any] = jnp.zeros(A ,dtype=jnp.floataa ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((1,) ,dtype=jnp.intaa ) UpperCAmelCase__ : int = jnp.zeros((1, 1, self.cross_attention_dim) ,dtype=jnp.floataa ) UpperCAmelCase__ : Dict = (1, 3, self.sample_size * 8, self.sample_size * 8) UpperCAmelCase__ : str = jnp.zeros(A ,dtype=jnp.floataa ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = jax.random.split(A ) UpperCAmelCase__ : Tuple = {"""params""": params_rng, """dropout""": dropout_rng} return self.init(A ,A ,A ,A ,A )["params"] def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.block_out_channels UpperCAmelCase__ : Dict = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. UpperCAmelCase__ : Dict = self.num_attention_heads or self.attention_head_dim # input UpperCAmelCase__ : Optional[Any] = nn.Conv( block_out_channels[0] ,kernel_size=(3, 3) ,strides=(1, 1) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) # time UpperCAmelCase__ : List[str] = FlaxTimesteps( block_out_channels[0] ,flip_sin_to_cos=self.flip_sin_to_cos ,freq_shift=self.config.freq_shift ) UpperCAmelCase__ : List[Any] = FlaxTimestepEmbedding(A ,dtype=self.dtype ) UpperCAmelCase__ : Dict = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] ,block_out_channels=self.conditioning_embedding_out_channels ,) UpperCAmelCase__ : Any = self.only_cross_attention if isinstance(A ,A ): UpperCAmelCase__ : Optional[int] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(A ,A ): UpperCAmelCase__ : Optional[Any] = (num_attention_heads,) * len(self.down_block_types ) # down UpperCAmelCase__ : Dict = [] UpperCAmelCase__ : Optional[int] = [] UpperCAmelCase__ : Dict = block_out_channels[0] UpperCAmelCase__ : Any = nn.Conv( A ,kernel_size=(1, 1) ,padding="""VALID""" ,kernel_init=nn.initializers.zeros_init() ,bias_init=nn.initializers.zeros_init() ,dtype=self.dtype ,) controlnet_down_blocks.append(A ) for i, down_block_type in enumerate(self.down_block_types ): UpperCAmelCase__ : Optional[Any] = output_channel UpperCAmelCase__ : Any = block_out_channels[i] UpperCAmelCase__ : List[str] = i == len(A ) - 1 if down_block_type == "CrossAttnDownBlock2D": UpperCAmelCase__ : Tuple = FlaxCrossAttnDownBlockaD( in_channels=A ,out_channels=A ,dropout=self.dropout ,num_layers=self.layers_per_block ,num_attention_heads=num_attention_heads[i] ,add_downsample=not is_final_block ,use_linear_projection=self.use_linear_projection ,only_cross_attention=only_cross_attention[i] ,dtype=self.dtype ,) else: UpperCAmelCase__ : List[str] = FlaxDownBlockaD( in_channels=A ,out_channels=A ,dropout=self.dropout ,num_layers=self.layers_per_block ,add_downsample=not is_final_block ,dtype=self.dtype ,) down_blocks.append(A ) for _ in range(self.layers_per_block ): UpperCAmelCase__ : Tuple = nn.Conv( A ,kernel_size=(1, 1) ,padding="""VALID""" ,kernel_init=nn.initializers.zeros_init() ,bias_init=nn.initializers.zeros_init() ,dtype=self.dtype ,) controlnet_down_blocks.append(A ) if not is_final_block: UpperCAmelCase__ : Optional[int] = nn.Conv( A ,kernel_size=(1, 1) ,padding="""VALID""" ,kernel_init=nn.initializers.zeros_init() ,bias_init=nn.initializers.zeros_init() ,dtype=self.dtype ,) controlnet_down_blocks.append(A ) UpperCAmelCase__ : str = down_blocks UpperCAmelCase__ : Optional[Any] = controlnet_down_blocks # mid UpperCAmelCase__ : Dict = block_out_channels[-1] UpperCAmelCase__ : Optional[Any] = FlaxUNetMidBlockaDCrossAttn( in_channels=A ,dropout=self.dropout ,num_attention_heads=num_attention_heads[-1] ,use_linear_projection=self.use_linear_projection ,dtype=self.dtype ,) UpperCAmelCase__ : Union[str, Any] = nn.Conv( A ,kernel_size=(1, 1) ,padding="""VALID""" ,kernel_init=nn.initializers.zeros_init() ,bias_init=nn.initializers.zeros_init() ,dtype=self.dtype ,) def __call__( self : List[str] ,A : Any ,A : Tuple ,A : str ,A : Optional[int] ,A : float = 1.0 ,A : bool = True ,A : bool = False ,): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.controlnet_conditioning_channel_order if channel_order == "bgr": UpperCAmelCase__ : Optional[Any] = jnp.flip(A ,axis=1 ) # 1. time if not isinstance(A ,jnp.ndarray ): UpperCAmelCase__ : Union[str, Any] = jnp.array([timesteps] ,dtype=jnp.intaa ) elif isinstance(A ,jnp.ndarray ) and len(timesteps.shape ) == 0: UpperCAmelCase__ : Dict = timesteps.astype(dtype=jnp.floataa ) UpperCAmelCase__ : Any = jnp.expand_dims(A ,0 ) UpperCAmelCase__ : str = self.time_proj(A ) UpperCAmelCase__ : List[Any] = self.time_embedding(A ) # 2. pre-process UpperCAmelCase__ : Dict = jnp.transpose(A ,(0, 2, 3, 1) ) UpperCAmelCase__ : Tuple = self.conv_in(A ) UpperCAmelCase__ : List[Any] = jnp.transpose(A ,(0, 2, 3, 1) ) UpperCAmelCase__ : Dict = self.controlnet_cond_embedding(A ) sample += controlnet_cond # 3. down UpperCAmelCase__ : str = (sample,) for down_block in self.down_blocks: if isinstance(A ,A ): UpperCAmelCase__ , UpperCAmelCase__ : int = down_block(A ,A ,A ,deterministic=not train ) else: UpperCAmelCase__ , UpperCAmelCase__ : str = down_block(A ,A ,deterministic=not train ) down_block_res_samples += res_samples # 4. mid UpperCAmelCase__ : Any = self.mid_block(A ,A ,A ,deterministic=not train ) # 5. contronet blocks UpperCAmelCase__ : Union[str, Any] = () for down_block_res_sample, controlnet_block in zip(A ,self.controlnet_down_blocks ): UpperCAmelCase__ : Tuple = controlnet_block(A ) controlnet_down_block_res_samples += (down_block_res_sample,) UpperCAmelCase__ : int = controlnet_down_block_res_samples UpperCAmelCase__ : Tuple = self.controlnet_mid_block(A ) # 6. scaling UpperCAmelCase__ : List[Any] = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=A ,mid_block_res_sample=A )
65
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece.model') __UpperCAmelCase = {'target_lang': 'fi', 'source_lang': 'en'} __UpperCAmelCase = '>>zh<<' __UpperCAmelCase = 'Helsinki-NLP/' if is_torch_available(): __UpperCAmelCase = 'pt' elif is_tf_available(): __UpperCAmelCase = 'tf' else: __UpperCAmelCase = 'jax' @require_sentencepiece class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = MarianTokenizer snake_case_ = False snake_case_ = True def __lowercase ( self : Optional[int] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Optional[Any] = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] UpperCAmelCase__ : int = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : Optional[int] = Path(self.tmpdirname ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(A ,save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(A ,save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) UpperCAmelCase__ : Dict = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self : List[Any] ,**A : List[Any] ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Union[str, Any] ,A : Tuple ): '''simple docstring''' return ( "This is a test", "This is a test", ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = """</s>""" UpperCAmelCase__ : int = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""</s>""" ) self.assertEqual(vocab_keys[1] ,"""<unk>""" ) self.assertEqual(vocab_keys[-1] ,"""<pad>""" ) self.assertEqual(len(A ) ,9 ) def __lowercase ( self : Dict ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,9 ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MarianTokenizer.from_pretrained(f"{ORG_NAME}opus-mt-en-de" ) UpperCAmelCase__ : List[str] = en_de_tokenizer(["""I am a small frog"""] ,return_tensors=A ) self.assertIsInstance(A ,A ) UpperCAmelCase__ : str = [38, 121, 14, 697, 38_848, 0] self.assertListEqual(A ,batch.input_ids[0] ) UpperCAmelCase__ : Optional[Any] = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(A ) UpperCAmelCase__ : Tuple = [x.name for x in Path(A ).glob("""*""" )] self.assertIn("""source.spm""" ,A ) MarianTokenizer.from_pretrained(A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Any = tok( ["""I am a small frog""" * 1_000, """I am a small frog"""] ,padding=A ,truncation=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch.input_ids.shape ,(2, 512) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizer() UpperCAmelCase__ : Tuple = tok(["""I am a tiny frog""", """I am a small frog"""] ,padding=A ,return_tensors=A ) self.assertIsInstance(A ,A ) self.assertEqual(batch_smaller.input_ids.shape ,(2, 10) ) @slow def __lowercase ( self : Dict ): '''simple docstring''' # fmt: off UpperCAmelCase__ : Optional[int] = {"""input_ids""": [[43_495, 462, 20, 42_164, 1_369, 52, 464, 132, 1_703, 492, 13, 7_491, 38_999, 6, 8, 464, 132, 1_703, 492, 13, 4_669, 37_867, 13, 7_525, 27, 1_593, 988, 13, 33_972, 7_029, 6, 20, 8_251, 383, 2, 270, 5_866, 3_788, 2, 2_353, 8_251, 12_338, 2, 13_958, 387, 2, 3_629, 6_953, 188, 2_900, 2, 13_958, 8_011, 11_501, 23, 8_460, 4_073, 34_009, 20, 435, 11_439, 27, 8, 8_460, 4_073, 6_004, 20, 9_988, 375, 27, 33, 266, 1_945, 1_076, 1_350, 37_867, 3_288, 5, 577, 1_076, 4_374, 8, 5_082, 5, 26_453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10_767, 6, 316, 304, 4_239, 3, 0], [148, 15_722, 19, 1_839, 12, 1_350, 13, 22_327, 5_082, 5_418, 47_567, 35_938, 59, 318, 19_552, 108, 2_183, 54, 14_976, 4_835, 32, 547, 1_114, 8, 315, 2_417, 5, 92, 19_088, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100], [36, 6_395, 12_570, 39_147, 11_597, 6, 266, 4, 45_405, 7_296, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""Helsinki-NLP/opus-mt-en-de""" ,revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" ,decode_kwargs={"""use_source_tokenizer""": True} ,) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[str] = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) UpperCAmelCase__ : Any = """Tämä on testi""" UpperCAmelCase__ : int = """This is a test""" UpperCAmelCase__ : List[str] = [76, 7, 2_047, 2] UpperCAmelCase__ : Optional[Any] = [69, 12, 11, 940, 2] UpperCAmelCase__ : List[str] = tokenizer(A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : Optional[int] = tokenizer(text_target=A ).input_ids self.assertListEqual(A ,A ) UpperCAmelCase__ : int = tokenizer.decode(A ,skip_special_tokens=A ) self.assertEqual(A ,A )
65
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class __lowercase ( __lowerCamelCase ): snake_case_ = """megatron-bert""" def __init__( self : Tuple ,A : Tuple=29_056 ,A : Dict=1_024 ,A : str=24 ,A : List[str]=16 ,A : List[str]=4_096 ,A : Any="gelu" ,A : Tuple=0.1 ,A : str=0.1 ,A : Dict=512 ,A : str=2 ,A : int=0.0_2 ,A : List[Any]=1e-12 ,A : Union[str, Any]=0 ,A : Union[str, Any]="absolute" ,A : Optional[int]=True ,**A : Optional[Any] ,): '''simple docstring''' super().__init__(pad_token_id=A ,**A ) UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : List[Any] = num_hidden_layers UpperCAmelCase__ : Optional[Any] = num_attention_heads UpperCAmelCase__ : List[Any] = hidden_act UpperCAmelCase__ : int = intermediate_size UpperCAmelCase__ : Tuple = hidden_dropout_prob UpperCAmelCase__ : Any = attention_probs_dropout_prob UpperCAmelCase__ : int = max_position_embeddings UpperCAmelCase__ : Optional[Any] = type_vocab_size UpperCAmelCase__ : Dict = initializer_range UpperCAmelCase__ : Optional[Any] = layer_norm_eps UpperCAmelCase__ : List[Any] = position_embedding_type UpperCAmelCase__ : Any = use_cache
65
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=__lowerCamelCase ): snake_case_ = ["""onnx"""] def __init__( self : int ,*A : List[str] ,**A : int ): '''simple docstring''' requires_backends(self ,["""onnx"""] ) @classmethod def __lowercase ( cls : Optional[Any] ,*A : List[str] ,**A : Dict ): '''simple docstring''' requires_backends(cls ,["""onnx"""] ) @classmethod def __lowercase ( cls : List[Any] ,*A : Optional[int] ,**A : int ): '''simple docstring''' requires_backends(cls ,["""onnx"""] )
65
1
"""simple docstring""" import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' monkeypatch.setattr("""datasets.utils.deprecation_utils._emitted_deprecation_warnings""" , set() ) @pytest.fixture def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' class __lowercase : def __init__( self : List[str] ,A : Tuple ): '''simple docstring''' UpperCAmelCase__ : Tuple = metric_id class __lowercase : snake_case_ = [MetricMock(__lowerCamelCase ) for metric_id in ["""accuracy""", """mse""", """precision""", """codeparrot/apps_metric"""]] def __lowercase ( self : List[Any] ): '''simple docstring''' return self._metrics monkeypatch.setattr("""datasets.inspect.huggingface_hub""" , HfhMock() ) @pytest.mark.parametrize( """func, args""" , [(load_metric, ("""metrics/mse""",)), (list_metrics, ()), (inspect_metric, ("""metrics/mse""", """tmp_path"""))] ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' if "tmp_path" in args: UpperCAmelCase__ : Any = tuple(arg if arg != """tmp_path""" else tmp_path for arg in args ) with pytest.warns(__UpperCamelCase , match="""https://huggingface.co/docs/evaluate""" ): func(*__UpperCamelCase )
65
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) UpperCAmelCase__ : List[Any] = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(__UpperCamelCase ) # Let's go UpperCAmelCase__ : int = parser.parse_args() if not hasattr(__UpperCamelCase , """func""" ): parser.print_help() exit(1 ) # Run UpperCAmelCase__ : Union[str, Any] = args.func(__UpperCamelCase ) service.run() if __name__ == "__main__": main()
65
1
"""simple docstring""" import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = WavaVecaPhonemeCTCTokenizer snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' super().setUp() UpperCAmelCase__ : Tuple = ( """<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː """ """ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː """ """ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 """ """oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ """ """pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ """ """yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ """ """əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ """ """ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ """ """ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ """ """uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ """ """ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ """ """ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ """ """ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4""" ).split(""" """ ) UpperCAmelCase__ : Dict = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase__ : List[Any] = {"""pad_token""": """<pad>""", """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>"""} UpperCAmelCase__ : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) def __lowercase ( self : Optional[Any] ,A : Dict ,A : Any=False ,A : List[str]=20 ,A : Tuple=5 ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [(i, tokenizer.decode([i] ,clean_up_tokenization_spaces=A )) for i in range(len(A ) )] UpperCAmelCase__ : Any = list(filter(lambda A : [t[0]] == tokenizer.encode(t[1] ,do_phonemize=A ) ,A ) ) if max_length is not None and len(A ) > max_length: UpperCAmelCase__ : Union[str, Any] = toks[:max_length] if min_length is not None and len(A ) < min_length and len(A ) > 0: while len(A ) < min_length: UpperCAmelCase__ : List[str] = toks + toks # toks_str = [t[1] for t in toks] UpperCAmelCase__ : Any = [t[0] for t in toks] # Ensure consistency UpperCAmelCase__ : Union[str, Any] = tokenizer.decode(A ,clean_up_tokenization_spaces=A ) if " " not in output_txt and len(A ) > 1: UpperCAmelCase__ : Optional[Any] = ( tokenizer.decode([toks_ids[0]] ,clean_up_tokenization_spaces=A ) + """ """ + tokenizer.decode(toks_ids[1:] ,clean_up_tokenization_spaces=A ) ) if with_prefix_space: UpperCAmelCase__ : Dict = """ """ + output_txt UpperCAmelCase__ : Any = tokenizer.encode(A ,add_special_tokens=A ) return output_txt, output_ids def __lowercase ( self : str ,**A : Dict ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname ,**A ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) # check adding a single token tokenizer.add_tokens("""xxx""" ) UpperCAmelCase__ : Optional[int] = tokenizer("""m xxx ɪ""" ,do_phonemize=A ).input_ids self.assertEqual(A ,[13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["""aaa""", """bbb""", """ccc"""] ) UpperCAmelCase__ : Optional[Any] = tokenizer("""m aaa ɪ ccc""" ,do_phonemize=A ).input_ids self.assertEqual(A ,[13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa UpperCAmelCase__ : Union[str, Any] = tokenizer("""maɪ c""" ,do_phonemize=A ).input_ids self.assertEqual(A ,[3, 200] ) # mai should be <unk> (=3) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Dict = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) UpperCAmelCase__ : int = """Hello how are you""" UpperCAmelCase__ : Any = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) self.assertEqual(A ,"""h ə l oʊ h aʊ ɑːɹ j uː""" ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) UpperCAmelCase__ : List[Any] = """Hello how are you""" UpperCAmelCase__ : Dict = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(A ).input_ids ,tokenizer(A ,do_phonemize=A ).input_ids ) def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) UpperCAmelCase__ : Optional[int] = """Hello how are you""" UpperCAmelCase__ : Optional[int] = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) UpperCAmelCase__ : List[str] = tokenizer.decode(tokenizer(A ).input_ids ) self.assertEqual(A ,A ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) UpperCAmelCase__ : List[str] = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] UpperCAmelCase__ : Optional[Any] = tokenizer.decode(sample_ids[0] ) UpperCAmelCase__ : str = tokenizer.batch_decode(A ) self.assertEqual(A ,batch_tokens[0] ) self.assertEqual(A ,["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" ,word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) UpperCAmelCase__ : Optional[Any] = """Hello how are you""" UpperCAmelCase__ : Optional[Any] = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) self.assertEqual(A ,"""h ə l oʊ | h aʊ | ɑːɹ | j uː |""" ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" ,word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) UpperCAmelCase__ : Tuple = """Hello how are you""" UpperCAmelCase__ : Tuple = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(A ).input_ids ,tokenizer(A ,do_phonemize=A ).input_ids ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" ,word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off UpperCAmelCase__ : Any = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter UpperCAmelCase__ : List[Any] = tokenizer.decode(sample_ids[0] ) UpperCAmelCase__ : str = tokenizer.batch_decode(A ) self.assertEqual(A ,batch_tokens[0] ) self.assertEqual(A ,["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) # decode with no word_del_token filter UpperCAmelCase__ : List[str] = tokenizer.decode(sample_ids[0] ,filter_word_delimiter_token=A ) UpperCAmelCase__ : int = tokenizer.batch_decode(A ,filter_word_delimiter_token=A ) self.assertEqual(A ,batch_tokens[0] ) self.assertEqual(A ,["""k s ɾ | ɾ l | ɭʲ""", """| j ð | s j ð s oːɹ"""] ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" ,word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) UpperCAmelCase__ : List[str] = """Hello how are you""" UpperCAmelCase__ : Any = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) UpperCAmelCase__ : str = tokenizer.decode(tokenizer(A ).input_ids ,filter_word_delimiter_token=A ) self.assertEqual(A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" ,word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) UpperCAmelCase__ : Tuple = """Hello how are you""" UpperCAmelCase__ : str = tokenizer.phonemize(A ,phonemizer_lang="""en-us""" ) UpperCAmelCase__ : Optional[Any] = tokenizer.decode(tokenizer(A ).input_ids ,filter_word_delimiter_token=A ) self.assertEqual(""" """.join([p.strip() for p in phonemes.split(""" |""" )] ).strip() ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" ,word_delimiter_token=A ) UpperCAmelCase__ : Optional[int] = """Hello how are you""" UpperCAmelCase__ : int = tokenizer(A ,phonemizer_lang="""en-us""" ).input_ids UpperCAmelCase__ : Optional[int] = tokenizer(A ,phonemizer_lang="""fr-fr""" ).input_ids self.assertNotEqual(A ,A ) UpperCAmelCase__ : List[Any] = tokenizer.decode(A ) UpperCAmelCase__ : Any = tokenizer.decode(A ) self.assertEqual(A ,"""h ə l oʊ h aʊ ɑːɹ j uː""" ) self.assertEqual(A ,"""ɛ l o h aʊ a ʁ j u""" ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) UpperCAmelCase__ : Dict = """Hello how Are you""" UpperCAmelCase__ : List[str] = """hello how are you""" UpperCAmelCase__ : Tuple = tokenizer(A ).input_ids UpperCAmelCase__ : Any = tokenizer(A ).input_ids self.assertEqual(A ,A ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) tokenizer.add_tokens(["""!""", """?"""] ) tokenizer.add_special_tokens({"""cls_token""": """$$$"""} ) # fmt: off UpperCAmelCase__ : Union[str, Any] = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on UpperCAmelCase__ : Union[str, Any] = tokenizer.batch_decode(A ) self.assertEqual(A ,["""k s ɾ ɾ l ɭʲ!?!? $$$""", """j ð s j ð s oːɹ $$$"""] ) @staticmethod def __lowercase ( A : Tuple ,A : str ): '''simple docstring''' UpperCAmelCase__ : Any = [d[key] for d in offsets] return retrieved_list def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer(word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" UpperCAmelCase__ : Union[str, Any] = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on UpperCAmelCase__ : Optional[Any] = tokenizer.decode(A ,output_char_offsets=A ,filter_word_delimiter_token=A ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) ,2 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""char_offsets""" in outputs ) self.assertTrue(isinstance(A ,A ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(""" """.join(self.get_from_offsets(outputs["""char_offsets"""] ,"""char""" ) ) ,outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] ,"""char""" ) ,["""k""", """s""", """ɾ""", """ɾ""", """|""", """ɾ""", """l""", """|""", """ɭʲ"""] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] ,"""start_offset""" ) ,[0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] ,"""end_offset""" ) ,[1, 4, 6, 9, 10, 12, 15, 16, 17] ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_tokenizer(word_delimiter_token="""|""" ) def check_list_tuples_equal(A : Union[str, Any] ,A : List[str] ): self.assertTrue(isinstance(A ,A ) ) self.assertTrue(isinstance(outputs_list[0] ,A ) ) # transform list to ModelOutput UpperCAmelCase__ : int = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["""text"""] ,outputs_batch_a["""text"""] ) def recursive_check(A : int ,A : Union[str, Any] ): if isinstance(A ,A ): [recursive_check(A ,A ) for la, la in zip(A ,A )] self.assertEqual(A ,A ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["""char_offsets"""] ,outputs_batch_a["""char_offsets"""] ) # fmt: off UpperCAmelCase__ : List[Any] = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char UpperCAmelCase__ : int = tokenizer.batch_decode(A ,output_char_offsets=A ) UpperCAmelCase__ : int = [tokenizer.decode(A ,output_char_offsets=A ) for ids in sample_ids] check_list_tuples_equal(A ,A ) @unittest.skip("""Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes""" ) def __lowercase ( self : List[str] ): '''simple docstring''' pass @unittest.skip("""Wav2Vec2PhonemeTokenizer always puts spaces between phonemes""" ) def __lowercase ( self : List[str] ): '''simple docstring''' pass @unittest.skip("""encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency""" ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip("""Wav2Vec2PhonemeModel has no max model length => no testing""" ) def __lowercase ( self : Tuple ): '''simple docstring''' pass def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : int = self.get_tokenizers(do_lower_case=A ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): UpperCAmelCase__ : Tuple = tokenizer.vocab_size UpperCAmelCase__ : List[Any] = len(A ) self.assertNotEqual(A ,0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) UpperCAmelCase__ : Dict = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] UpperCAmelCase__ : Optional[int] = tokenizer.add_tokens(A ) UpperCAmelCase__ : int = tokenizer.vocab_size UpperCAmelCase__ : Dict = len(A ) self.assertNotEqual(A ,0 ) self.assertEqual(A ,A ) self.assertEqual(A ,len(A ) ) self.assertEqual(A ,all_size + len(A ) ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" ,add_special_tokens=A ) self.assertGreaterEqual(len(A ) ,4 ) self.assertGreater(tokens[0] ,tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] ,tokenizer.vocab_size - 1 ) UpperCAmelCase__ : List[Any] = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} UpperCAmelCase__ : List[str] = tokenizer.add_special_tokens(A ) UpperCAmelCase__ : Optional[int] = tokenizer.vocab_size UpperCAmelCase__ : Union[str, Any] = len(A ) self.assertNotEqual(A ,0 ) self.assertEqual(A ,A ) self.assertEqual(A ,len(A ) ) self.assertEqual(A ,all_size_a + len(A ) ) UpperCAmelCase__ : Union[str, Any] = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" ,add_special_tokens=A ) self.assertGreaterEqual(len(A ) ,6 ) self.assertGreater(tokens[0] ,tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] ,tokens[1] ) self.assertGreater(tokens[-3] ,tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] ,tokens[-4] ) self.assertEqual(tokens[0] ,tokenizer.eos_token_id ) self.assertEqual(tokens[-3] ,tokenizer.pad_token_id ) @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def __lowercase ( self : Tuple ): '''simple docstring''' pass @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def __lowercase ( self : Optional[int] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' # The default common tokenizer tests assumes that the output of `convert_tokens_to_string` is a string which # is not the case for Wav2Vec2PhonemeCTCTokenizer. UpperCAmelCase__ : List[str] = self.get_tokenizers(fast=A ,do_lower_case=A ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): UpperCAmelCase__ : str = ["""ð""", """ɪ""", """s""", """ɪ""", """z""", """ɐ""", """t""", """ɛ""", """k""", """s""", """t"""] UpperCAmelCase__ : int = tokenizer.convert_tokens_to_string(A ) self.assertIsInstance(output["""text"""] ,A )
65
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __UpperCAmelCase = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __lowercase : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : List[Any] ,A : int ,A : Optional[Any]=13 ,A : Dict=7 ,A : Dict=True ,A : Any=False ,A : Dict=99 ,A : int=32 ,A : Optional[int]=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : str=0.1 ,A : int=0.1 ,A : Optional[int]=20 ,A : Tuple=2 ,A : str=1 ,A : Optional[Any]=0 ,): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : List[Any] = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : int = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Any = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : str = max_position_embeddings UpperCAmelCase__ : Union[str, Any] = eos_token_id UpperCAmelCase__ : Union[str, Any] = pad_token_id UpperCAmelCase__ : List[str] = bos_token_id def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length - 1] ,self.vocab_size ).clip(3 ,self.vocab_size ) UpperCAmelCase__ : List[str] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) ,1 ) UpperCAmelCase__ : Any = np.concatenate([input_ids, eos_tensor] ,axis=1 ) UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase__ : str = self.config_cls( vocab_size=self.vocab_size ,d_model=self.hidden_size ,encoder_layers=self.num_hidden_layers ,decoder_layers=self.num_hidden_layers ,encoder_attention_heads=self.num_attention_heads ,decoder_attention_heads=self.num_attention_heads ,encoder_ffn_dim=self.intermediate_size ,decoder_ffn_dim=self.intermediate_size ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,eos_token_ids=[2] ,bos_token_id=self.bos_token_id ,pad_token_id=self.pad_token_id ,decoder_start_token_id=self.pad_token_id ,**self.config_updates ,) UpperCAmelCase__ : Optional[Any] = prepare_pegasus_inputs_dict(A ,A ,A ) return config, inputs_dict def __lowercase ( self : Any ,A : Optional[int] ,A : str ,A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : Dict = model_class_name(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : Union[str, Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) ,dtype="""i4""" ) UpperCAmelCase__ : str = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Optional[int] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : int = model.decode( decoder_input_ids[:, -1:] ,A ,decoder_attention_mask=A ,past_key_values=outputs_cache.past_key_values ,decoder_position_ids=A ,) UpperCAmelCase__ : Dict = model.decode(A ,A ) UpperCAmelCase__ : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def __lowercase ( self : Optional[int] ,A : str ,A : Dict ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = 20 UpperCAmelCase__ : str = model_class_name(A ) UpperCAmelCase__ : Any = model.encode(inputs_dict["""input_ids"""] ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) UpperCAmelCase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] ,axis=-1 ,) UpperCAmelCase__ : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] ,A ,A ) UpperCAmelCase__ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] ,(decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) ,) UpperCAmelCase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] ,A ,decoder_attention_mask=A ,past_key_values=A ,decoder_position_ids=A ,) UpperCAmelCase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] ,dtype="""i4""" ) UpperCAmelCase__ : Dict = model.decode( decoder_input_ids[:, -1:] ,A ,past_key_values=outputs_cache.past_key_values ,decoder_attention_mask=A ,decoder_position_ids=A ,) UpperCAmelCase__ : Union[str, Any] = model.decode(A ,A ,decoder_attention_mask=A ) UpperCAmelCase__ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 ,msg=f"Max diff is {diff}" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , ): '''simple docstring''' if attention_mask is None: UpperCAmelCase__ : Union[str, Any] = np.not_equal(__UpperCamelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: UpperCAmelCase__ : Tuple = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) snake_case_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = FlaxPegasusModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self ,config_class=A ) def __lowercase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(A ,A ,A ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(A ,A ,A ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : List[Any] = self._prepare_for_class(A ,A ) UpperCAmelCase__ : int = model_class(A ) @jax.jit def encode_jitted(A : Optional[int] ,A : Union[str, Any]=None ,**A : Optional[Any] ): return model.encode(input_ids=A ,attention_mask=A ) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : int = encode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : Dict = encode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase__ : Dict = model_class(A ) UpperCAmelCase__ : str = model.encode(inputs_dict["""input_ids"""] ,inputs_dict["""attention_mask"""] ) UpperCAmelCase__ : Dict = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(A : List[Any] ,A : Any ,A : List[Any] ): return model.decode( decoder_input_ids=A ,decoder_attention_mask=A ,encoder_outputs=A ,) with self.subTest("""JIT Enabled""" ): UpperCAmelCase__ : Tuple = decode_jitted(**A ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): UpperCAmelCase__ : str = decode_jitted(**A ).to_tuple() self.assertEqual(len(A ) ,len(A ) ) for jitted_output, output in zip(A ,A ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""google/pegasus-large""" ,from_pt=A ) UpperCAmelCase__ : Any = np.ones((1, 1) ) UpperCAmelCase__ : Optional[Any] = model(A ) self.assertIsNotNone(A ) @slow def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Dict = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Optional[Any] = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""" ) UpperCAmelCase__ : Union[str, Any] = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] UpperCAmelCase__ : str = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] UpperCAmelCase__ : str = tokenizer(A ,return_tensors="""np""" ,truncation=A ,max_length=512 ,padding=A ) UpperCAmelCase__ : Union[str, Any] = model.generate(**A ,num_beams=2 ).sequences UpperCAmelCase__ : int = tokenizer.batch_decode(A ,skip_special_tokens=A ) assert tgt_text == decoded
65
1
"""simple docstring""" from math import factorial def lowerCAmelCase ( __UpperCamelCase = 100 ): '''simple docstring''' return sum(int(__UpperCamelCase ) for x in str(factorial(__UpperCamelCase ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
65
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) UpperCAmelCase__ : Union[str, Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
65
1
"""simple docstring""" def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) UpperCAmelCase__ : Union[str, Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
65
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase ( __UpperCamelCase = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCAmelCase__ : Dict = F"{olid} is not a valid Open Library olid" raise ValueError(__UpperCamelCase ) return requests.get(F"https://openlibrary.org/{new_olid}.json" ).json() def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } UpperCAmelCase__ : Dict = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase__ : str = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] UpperCAmelCase__ : Dict = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase__ : Dict = """, """.join(__UpperCamelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __UpperCAmelCase = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.") continue print(F"\nSearching Open Library for ISBN: {isbn}...\n") try: __UpperCAmelCase = summarize_book(get_openlibrary_data(F"isbn/{isbn}")) print('\n'.join(F"{key}: {value}" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"Sorry, there are no results for ISBN: {isbn}.")
65
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: __UpperCAmelCase = None __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} __UpperCAmelCase = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/tokenizer.json', }, } __UpperCAmelCase = { 'camembert-base': 512, } __UpperCAmelCase = '▁' class __lowercase ( __lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ["""input_ids""", """attention_mask"""] snake_case_ = CamembertTokenizer def __init__( self : List[str] ,A : Optional[int]=None ,A : List[str]=None ,A : List[Any]="<s>" ,A : Optional[int]="</s>" ,A : Optional[Any]="</s>" ,A : str="<s>" ,A : Optional[Any]="<unk>" ,A : Tuple="<pad>" ,A : int="<mask>" ,A : Tuple=["<s>NOTUSED", "</s>NOTUSED"] ,**A : str ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ : List[str] = AddedToken(A ,lstrip=A ,rstrip=A ) if isinstance(A ,A ) else mask_token super().__init__( A ,tokenizer_file=A ,bos_token=A ,eos_token=A ,sep_token=A ,cls_token=A ,unk_token=A ,pad_token=A ,mask_token=A ,additional_special_tokens=A ,**A ,) UpperCAmelCase__ : List[str] = vocab_file UpperCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def __lowercase ( self : Tuple ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] UpperCAmelCase__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : str ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __lowercase ( self : str ,A : str ,A : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(A ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase__ : Dict = os.path.join( A ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ): copyfile(self.vocab_file ,A ) return (out_vocab_file,)
65
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=True , __UpperCamelCase="pt" ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {"""add_prefix_space""": True} if isinstance(__UpperCamelCase , __UpperCamelCase ) and not line.startswith(""" """ ) else {} UpperCAmelCase__ : List[str] = padding_side return tokenizer( [line] , max_length=__UpperCamelCase , padding="""max_length""" if pad_to_max_length else None , truncation=__UpperCamelCase , return_tensors=__UpperCamelCase , add_special_tokens=__UpperCamelCase , **__UpperCamelCase , ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , ): '''simple docstring''' UpperCAmelCase__ : str = input_ids.ne(__UpperCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __lowercase ( __lowerCamelCase ): def __init__( self : Tuple ,A : List[Any] ,A : Union[str, Any] ,A : Any ,A : Optional[int] ,A : Union[str, Any]="train" ,A : Tuple=None ,A : Union[str, Any]=None ,A : Tuple=None ,A : int="" ,): '''simple docstring''' super().__init__() UpperCAmelCase__ : Optional[Any] = Path(A ).joinpath(type_path + """.source""" ) UpperCAmelCase__ : List[str] = Path(A ).joinpath(type_path + """.target""" ) UpperCAmelCase__ : Dict = self.get_char_lens(self.src_file ) UpperCAmelCase__ : int = max_source_length UpperCAmelCase__ : List[str] = max_target_length assert min(self.src_lens ) > 0, f"found empty line in {self.src_file}" UpperCAmelCase__ : Dict = tokenizer UpperCAmelCase__ : str = prefix if n_obs is not None: UpperCAmelCase__ : int = self.src_lens[:n_obs] UpperCAmelCase__ : Any = src_lang UpperCAmelCase__ : Any = tgt_lang def __len__( self : Optional[Any] ): '''simple docstring''' return len(self.src_lens ) def __getitem__( self : Union[str, Any] ,A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = index + 1 # linecache starts at 1 UpperCAmelCase__ : Tuple = self.prefix + linecache.getline(str(self.src_file ) ,A ).rstrip("""\n""" ) UpperCAmelCase__ : Dict = linecache.getline(str(self.tgt_file ) ,A ).rstrip("""\n""" ) assert source_line, f"empty source line for index {index}" assert tgt_line, f"empty tgt line for index {index}" # Need to add eos token manually for T5 if isinstance(self.tokenizer ,A ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right UpperCAmelCase__ : str = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,A ) else self.tokenizer ) UpperCAmelCase__ : Tuple = self.tokenizer.generator if isinstance(self.tokenizer ,A ) else self.tokenizer UpperCAmelCase__ : Tuple = encode_line(A ,A ,self.max_source_length ,"""right""" ) UpperCAmelCase__ : Dict = encode_line(A ,A ,self.max_target_length ,"""right""" ) UpperCAmelCase__ : Optional[Any] = source_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : List[str] = target_inputs["""input_ids"""].squeeze() UpperCAmelCase__ : Union[str, Any] = source_inputs["""attention_mask"""].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def __lowercase ( A : int ): '''simple docstring''' return [len(A ) for x in Path(A ).open().readlines()] def __lowercase ( self : List[Any] ,A : Any ): '''simple docstring''' UpperCAmelCase__ : int = torch.stack([x["""input_ids"""] for x in batch] ) UpperCAmelCase__ : Union[str, Any] = torch.stack([x["""attention_mask"""] for x in batch] ) UpperCAmelCase__ : Any = torch.stack([x["""decoder_input_ids"""] for x in batch] ) UpperCAmelCase__ : List[Any] = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : Any = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,A ) else self.tokenizer.pad_token_id ) UpperCAmelCase__ : str = trim_batch(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = trim_batch(A ,A ,attention_mask=A ) UpperCAmelCase__ : List[str] = { """input_ids""": source_ids, """attention_mask""": source_mask, """decoder_input_ids""": y, } return batch __UpperCAmelCase = getLogger(__name__) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return list(itertools.chain.from_iterable(__UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = get_git_info() save_json(__UpperCamelCase , os.path.join(__UpperCamelCase , """git_log.json""" ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=4 , **__UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=__UpperCamelCase , **__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = git.Repo(search_parent_directories=__UpperCamelCase ) UpperCAmelCase__ : List[str] = { """repo_id""": str(__UpperCamelCase ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), """hostname""": str(socket.gethostname() ), } return repo_infos def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return list(map(__UpperCamelCase , __UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """wb""" ) as f: return pickle.dump(__UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' def remove_articles(__UpperCamelCase ): return re.sub(r"""\b(a|an|the)\b""" , """ """ , __UpperCamelCase ) def white_space_fix(__UpperCamelCase ): return " ".join(text.split() ) def remove_punc(__UpperCamelCase ): UpperCAmelCase__ : List[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCamelCase ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCamelCase ) ) ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : Dict = normalize_answer(__UpperCamelCase ).split() UpperCAmelCase__ : int = Counter(__UpperCamelCase ) & Counter(__UpperCamelCase ) UpperCAmelCase__ : List[str] = sum(common.values() ) if num_same == 0: return 0 UpperCAmelCase__ : str = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = 1.0 * num_same / len(__UpperCamelCase ) UpperCAmelCase__ : Tuple = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return normalize_answer(__UpperCamelCase ) == normalize_answer(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' assert len(__UpperCamelCase ) == len(__UpperCamelCase ) UpperCAmelCase__ : Union[str, Any] = 0 for hypo, pred in zip(__UpperCamelCase , __UpperCamelCase ): em += exact_match_score(__UpperCamelCase , __UpperCamelCase ) if len(__UpperCamelCase ) > 0: em /= len(__UpperCamelCase ) return {"em": em} def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' return model_prefix.startswith("""rag""" ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead UpperCAmelCase__ : str = """dropout_rate""" for p in extra_params: if getattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): if not hasattr(__UpperCamelCase , __UpperCamelCase ) and not hasattr(__UpperCamelCase , equivalent_param[p] ): logger.info("""config doesn't have a `{}` attribute""".format(__UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) continue UpperCAmelCase__ : Tuple = p if hasattr(__UpperCamelCase , __UpperCamelCase ) else equivalent_param[p] setattr(__UpperCamelCase , __UpperCamelCase , getattr(__UpperCamelCase , __UpperCamelCase ) ) delattr(__UpperCamelCase , __UpperCamelCase ) return hparams, config
65
1
"""simple docstring""" import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def lowerCAmelCase ( __UpperCamelCase ): # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def lowerCAmelCase ( ): '''simple docstring''' with parallel_backend("""spark""" ): assert ParallelBackendConfig.backend_name == "spark" UpperCAmelCase__ : int = [1, 2, 3] with pytest.raises(__UpperCamelCase ): with parallel_backend("""unsupported backend""" ): map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=2 ) with pytest.raises(__UpperCamelCase ): with parallel_backend("""unsupported backend""" ): map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize("""num_proc""" , [2, -1] ) def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = [1, 2] UpperCAmelCase__ : int = {"""a""": 1, """b""": 2} UpperCAmelCase__ : int = {"""a""": [1, 2], """b""": [3, 4]} UpperCAmelCase__ : Optional[int] = {"""a""": {"""1""": 1}, """b""": 2} UpperCAmelCase__ : Dict = {"""a""": 1, """b""": 2, """c""": 3, """d""": 4} UpperCAmelCase__ : Tuple = [2, 3] UpperCAmelCase__ : Tuple = {"""a""": 2, """b""": 3} UpperCAmelCase__ : int = {"""a""": [2, 3], """b""": [4, 5]} UpperCAmelCase__ : Tuple = {"""a""": {"""1""": 2}, """b""": 3} UpperCAmelCase__ : Union[str, Any] = {"""a""": 2, """b""": 3, """c""": 4, """d""": 5} with parallel_backend("""spark""" ): assert map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=__UpperCamelCase ) == expected_map_nested_sa assert map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=__UpperCamelCase ) == expected_map_nested_sa assert map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=__UpperCamelCase ) == expected_map_nested_sa assert map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=__UpperCamelCase ) == expected_map_nested_sa assert map_nested(__UpperCamelCase , __UpperCamelCase , num_proc=__UpperCamelCase ) == expected_map_nested_sa
65
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowercase ( __lowerCamelCase , unittest.TestCase ): snake_case_ = KandinskyVaaControlnetPipeline snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case_ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case_ = False @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return 32 @property def __lowercase ( self : int ): '''simple docstring''' return 32 @property def __lowercase ( self : Dict ): '''simple docstring''' return self.time_input_dim @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self : Any ): '''simple docstring''' return 100 @property def __lowercase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : Tuple = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } UpperCAmelCase__ : int = UNetaDConditionModel(**A ) return model @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase__ : str = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : str = self.dummy_unet UpperCAmelCase__ : List[Any] = self.dummy_movq UpperCAmelCase__ : List[Any] = DDIMScheduler( num_train_timesteps=1_000 ,beta_schedule="""linear""" ,beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,clip_sample=A ,set_alpha_to_one=A ,steps_offset=1 ,prediction_type="""epsilon""" ,thresholding=A ,) UpperCAmelCase__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __lowercase ( self : str ,A : Optional[Any] ,A : Any=0 ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(A ) ).to(A ) UpperCAmelCase__ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( A ) # create hint UpperCAmelCase__ : int = floats_tensor((1, 3, 64, 64) ,rng=random.Random(A ) ).to(A ) if str(A ).startswith("""mps""" ): UpperCAmelCase__ : Optional[int] = torch.manual_seed(A ) else: UpperCAmelCase__ : Dict = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase__ : Dict = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Dict = """cpu""" UpperCAmelCase__ : List[Any] = self.get_dummy_components() UpperCAmelCase__ : Union[str, Any] = self.pipeline_class(**A ) UpperCAmelCase__ : Optional[int] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[int] = pipe(**self.get_dummy_inputs(A ) ) UpperCAmelCase__ : Tuple = output.images UpperCAmelCase__ : Dict = pipe( **self.get_dummy_inputs(A ) ,return_dict=A ,)[0] UpperCAmelCase__ : Tuple = image[0, -3:, -3:, -1] UpperCAmelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase__ : Optional[int] = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def __lowercase ( self : Union[str, Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) UpperCAmelCase__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) UpperCAmelCase__ : int = torch.from_numpy(np.array(A ) ).float() / 2_5_5.0 UpperCAmelCase__ : Union[str, Any] = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) UpperCAmelCase__ : List[str] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" ,torch_dtype=torch.floataa ) pipe_prior.to(A ) UpperCAmelCase__ : List[Any] = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" ,torch_dtype=torch.floataa ) UpperCAmelCase__ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) UpperCAmelCase__ : Optional[Any] = """A robot, 4k photo""" UpperCAmelCase__ : List[Any] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ , UpperCAmelCase__ : Tuple = pipe_prior( A ,generator=A ,num_inference_steps=5 ,negative_prompt="""""" ,).to_tuple() UpperCAmelCase__ : List[str] = torch.Generator(device="""cuda""" ).manual_seed(0 ) UpperCAmelCase__ : int = pipeline( image_embeds=A ,negative_image_embeds=A ,hint=A ,generator=A ,num_inference_steps=100 ,output_type="""np""" ,) UpperCAmelCase__ : Any = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A ,A )
65
1
"""simple docstring""" import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = TaConfig.from_json_file(__UpperCamelCase ) print(F"Building PyTorch model from configuration: {config}" ) UpperCAmelCase__ : Any = TaForConditionalGeneration(__UpperCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_ta(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __UpperCAmelCase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
65
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __UpperCAmelCase = logging.get_logger(__name__) class __lowercase ( __lowerCamelCase ): snake_case_ = """vision-encoder-decoder""" snake_case_ = True def __init__( self : List[Any] ,**A : Union[str, Any] ): '''simple docstring''' super().__init__(**A ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) UpperCAmelCase__ : int = kwargs.pop("""encoder""" ) UpperCAmelCase__ : int = encoder_config.pop("""model_type""" ) UpperCAmelCase__ : str = kwargs.pop("""decoder""" ) UpperCAmelCase__ : Dict = decoder_config.pop("""model_type""" ) UpperCAmelCase__ : List[Any] = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Any = AutoConfig.for_model(A ,**A ) UpperCAmelCase__ : Union[str, Any] = True @classmethod def __lowercase ( cls : List[Any] ,A : PretrainedConfig ,A : PretrainedConfig ,**A : Tuple ): '''simple docstring''' logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : List[Any] = True return cls(encoder=encoder_config.to_dict() ,decoder=decoder_config.to_dict() ,**A ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ : Dict = self.encoder.to_dict() UpperCAmelCase__ : Any = self.decoder.to_dict() UpperCAmelCase__ : Dict = self.__class__.model_type return output class __lowercase ( __lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __lowercase ( self : List[Any] ): '''simple docstring''' return 1e-4 @property def __lowercase ( self : List[Any] ): '''simple docstring''' return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : Optional[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCAmelCase__ : List[str] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def __lowercase ( self : Dict ,A : "PreTrainedTokenizerBase" ,A : int = -1 ,A : int = -1 ,A : bool = False ,A : Optional["TensorType"] = None ,): '''simple docstring''' import torch UpperCAmelCase__ : int = OrderedDict() UpperCAmelCase__ : List[Any] = super().generate_dummy_inputs( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) UpperCAmelCase__ , UpperCAmelCase__ : int = dummy_input["""input_ids"""].shape UpperCAmelCase__ : int = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCAmelCase__ : Tuple = dummy_input.pop("""input_ids""" ) UpperCAmelCase__ : Optional[int] = dummy_input.pop("""attention_mask""" ) UpperCAmelCase__ : Dict = torch.zeros(A ) return common_inputs class __lowercase ( __lowerCamelCase ): @property def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : Any ,A : PretrainedConfig ): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(A ) def __lowercase ( self : Dict ,A : PretrainedConfig ,A : PretrainedConfig ,A : str = "default" ): '''simple docstring''' UpperCAmelCase__ : List[str] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(A ,A )
65
1