code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def _UpperCAmelCase ( self : Optional[int] ): _a = tempfile.mkdtemp() # fmt: off _a = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on _a = dict(zip(_A , range(len(_A ) ) ) ) _a = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>''', ''''''] _a = {'''unk_token''': '''<unk>'''} _a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) _a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_A ) ) _a = { '''do_resize''': True, '''size''': 2_0, '''do_center_crop''': True, '''crop_size''': 1_8, '''do_normalize''': True, '''image_mean''': [0.48145466, 0.4578275, 0.40821073], '''image_std''': [0.26862954, 0.26130258, 0.27577711], } _a = os.path.join(self.tmpdirname , _A ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_A , _A ) def _UpperCAmelCase ( self : Optional[Any] , **SCREAMING_SNAKE_CASE_ : List[str] ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **_A ) def _UpperCAmelCase ( self : Any , **SCREAMING_SNAKE_CASE_ : Dict ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **_A ) def _UpperCAmelCase ( self : Union[str, Any] , **SCREAMING_SNAKE_CASE_ : int ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **_A ) def _UpperCAmelCase ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def _UpperCAmelCase ( self : Optional[Any] ): _a = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] _a = [Image.fromarray(np.moveaxis(_A , 0 , -1 ) ) for x in image_inputs] return image_inputs def _UpperCAmelCase ( self : str ): _a = self.get_tokenizer() _a = self.get_rust_tokenizer() _a = self.get_image_processor() _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) processor_slow.save_pretrained(self.tmpdirname ) _a = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=_A ) _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) processor_fast.save_pretrained(self.tmpdirname ) _a = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _A ) self.assertIsInstance(processor_fast.tokenizer , _A ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _A ) self.assertIsInstance(processor_fast.image_processor , _A ) def _UpperCAmelCase ( self : List[Any] ): _a = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) _a = self.get_image_processor(do_normalize=_A , padding_value=1.0 ) _a = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def _UpperCAmelCase ( self : List[str] ): _a = self.get_image_processor() _a = self.get_tokenizer() _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) _a = self.prepare_image_inputs() _a = image_processor(_A , return_tensors='np' ) _a = processor(images=_A , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _UpperCAmelCase ( self : int ): _a = self.get_image_processor() _a = self.get_tokenizer() _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) _a = '''lower newer''' _a = processor(text=_A ) _a = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _UpperCAmelCase ( self : List[str] ): _a = self.get_image_processor() _a = self.get_tokenizer() _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) _a = '''lower newer''' _a = self.prepare_image_inputs() _a = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def _UpperCAmelCase ( self : int ): _a = self.get_image_processor() _a = self.get_tokenizer() _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) _a = self.prepare_image_inputs() _a = self.prepare_image_inputs() _a = processor(images=_A , visual_prompt=_A ) self.assertListEqual(list(inputs.keys() ) , ['pixel_values', 'conditional_pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def _UpperCAmelCase ( self : Union[str, Any] ): _a = self.get_image_processor() _a = self.get_tokenizer() _a = CLIPSegProcessor(tokenizer=_A , image_processor=_A ) _a = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a = processor.batch_decode(_A ) _a = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A )
562
def _lowerCamelCase ( __A : float ) -> float: if edge <= 0 or not isinstance(__A , __A ): raise ValueError('''Length must be a positive.''' ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def _lowerCamelCase ( __A : float ) -> float: if edge <= 0 or not isinstance(__A , __A ): raise ValueError('''Length must be a positive.''' ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
485
0
from __future__ import annotations __UpperCamelCase : Optional[Any] = list[tuple[int, int]] __UpperCamelCase : Optional[Any] = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __UpperCamelCase : Optional[Any] = ([-1, 0], [0, -1], [1, 0], [0, 1]) # up, left, down, right class _UpperCamelCase : '''simple docstring''' def __init__( self : Optional[Any] , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : Node | None , ): '''simple docstring''' __lowerCamelCase : Dict = pos_x __lowerCamelCase : List[str] = pos_y __lowerCamelCase : List[Any] = (pos_y, pos_x) __lowerCamelCase : List[Any] = goal_x __lowerCamelCase : Tuple = goal_y __lowerCamelCase : Optional[int] = g_cost __lowerCamelCase : Tuple = parent __lowerCamelCase : int = self.calculate_heuristic() def _snake_case ( self : Optional[int] ): '''simple docstring''' __lowerCamelCase : str = abs(self.pos_x - self.goal_x ) __lowerCamelCase : Optional[int] = abs(self.pos_y - self.goal_y ) return dx + dy def __lt__( self : Optional[int] , _lowerCamelCase : Tuple ): '''simple docstring''' return self.f_cost < other.f_cost class _UpperCamelCase : '''simple docstring''' def __init__( self : Any , _lowerCamelCase : tuple[int, int] , _lowerCamelCase : tuple[int, int] ): '''simple docstring''' __lowerCamelCase : List[Any] = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , _lowerCamelCase ) __lowerCamelCase : Optional[int] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9_9_9_9 , _lowerCamelCase ) __lowerCamelCase : int = [self.start] __lowerCamelCase : list[Node] = [] __lowerCamelCase : List[str] = False def _snake_case ( self : str ): '''simple docstring''' while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() __lowerCamelCase : Optional[Any] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: __lowerCamelCase : Union[str, Any] = True return self.retrace_path(_lowerCamelCase ) self.closed_nodes.append(_lowerCamelCase ) __lowerCamelCase : Any = self.get_successors(_lowerCamelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(_lowerCamelCase ) else: # retrieve the best current path __lowerCamelCase : int = self.open_nodes.pop(self.open_nodes.index(_lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(_lowerCamelCase ) else: self.open_nodes.append(_lowerCamelCase ) if not self.reached: return [self.start.pos] return None def _snake_case ( self : str , _lowerCamelCase : Node ): '''simple docstring''' __lowerCamelCase : str = [] for action in delta: __lowerCamelCase : List[str] = parent.pos_x + action[1] __lowerCamelCase : Tuple = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(_lowerCamelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( _lowerCamelCase , _lowerCamelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , _lowerCamelCase , ) ) return successors def _snake_case ( self : Dict , _lowerCamelCase : Node | None ): '''simple docstring''' __lowerCamelCase : Union[str, Any] = node __lowerCamelCase : Dict = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) __lowerCamelCase : Dict = current_node.parent path.reverse() return path if __name__ == "__main__": __UpperCamelCase : Optional[int] = (0, 0) __UpperCamelCase : Any = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) print('------') __UpperCamelCase : Optional[int] = GreedyBestFirst(init, goal) __UpperCamelCase : List[Any] = greedy_bf.search() if path: for pos_x, pos_y in path: __UpperCamelCase : Optional[Any] = 2 for elem in grid: print(elem)
458
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _UpperCamelCase ( A ): '''simple docstring''' def __init__( self : str , _lowerCamelCase : str = "▁" , _lowerCamelCase : bool = True , _lowerCamelCase : Union[str, AddedToken] = "<unk>" , _lowerCamelCase : Union[str, AddedToken] = "</s>" , _lowerCamelCase : Union[str, AddedToken] = "<pad>" , ): '''simple docstring''' __lowerCamelCase : Tuple = { """pad""": {"""id""": 0, """token""": pad_token}, """eos""": {"""id""": 1, """token""": eos_token}, """unk""": {"""id""": 2, """token""": unk_token}, } __lowerCamelCase : Optional[Any] = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): __lowerCamelCase : Tuple = token_dict["""token"""] __lowerCamelCase : Optional[int] = Tokenizer(Unigram() ) __lowerCamelCase : str = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(""" {2,}""" ) , """ """ ), normalizers.Lowercase(), ] ) __lowerCamelCase : int = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=_lowerCamelCase , add_prefix_space=_lowerCamelCase ), pre_tokenizers.Digits(individual_digits=_lowerCamelCase ), pre_tokenizers.Punctuation(), ] ) __lowerCamelCase : Tuple = decoders.Metaspace(replacement=_lowerCamelCase , add_prefix_space=_lowerCamelCase ) __lowerCamelCase : Tuple = TemplateProcessing( single=F"""$A {self.special_tokens["eos"]["token"]}""" , special_tokens=[(self.special_tokens["""eos"""]["""token"""], self.special_tokens["""eos"""]["""id"""])] , ) __lowerCamelCase : Tuple = { """model""": """SentencePieceUnigram""", """replacement""": replacement, """add_prefix_space""": add_prefix_space, } super().__init__(_lowerCamelCase , _lowerCamelCase ) def _snake_case ( self : Dict , _lowerCamelCase : Union[str, List[str]] , _lowerCamelCase : int = 8_0_0_0 , _lowerCamelCase : bool = True , ): '''simple docstring''' __lowerCamelCase : Optional[Any] = trainers.UnigramTrainer( vocab_size=_lowerCamelCase , special_tokens=self.special_tokens_list , show_progress=_lowerCamelCase , ) if isinstance(_lowerCamelCase , _lowerCamelCase ): __lowerCamelCase : Optional[int] = [files] self._tokenizer.train(_lowerCamelCase , trainer=_lowerCamelCase ) self.add_unk_id() def _snake_case ( self : List[Any] , _lowerCamelCase : Union[Iterator[str], Iterator[Iterator[str]]] , _lowerCamelCase : int = 8_0_0_0 , _lowerCamelCase : bool = True , ): '''simple docstring''' __lowerCamelCase : Any = trainers.UnigramTrainer( vocab_size=_lowerCamelCase , special_tokens=self.special_tokens_list , show_progress=_lowerCamelCase , ) self._tokenizer.train_from_iterator(_lowerCamelCase , trainer=_lowerCamelCase ) self.add_unk_id() def _snake_case ( self : List[str] ): '''simple docstring''' __lowerCamelCase : List[str] = json.loads(self._tokenizer.to_str() ) __lowerCamelCase : Optional[int] = self.special_tokens["""unk"""]["""id"""] __lowerCamelCase : List[str] = Tokenizer.from_str(json.dumps(_lowerCamelCase ) )
458
1
from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCamelCase_ : '''simple docstring''' def __init__( self : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any]=3 , UpperCAmelCase__ : Optional[Any]=32 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : List[Any]=10 , UpperCAmelCase__ : int=[10, 20, 30, 40] , UpperCAmelCase__ : str=[1, 1, 2, 1] , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : List[str]="relu" , UpperCAmelCase__ : Optional[int]=3 , UpperCAmelCase__ : Tuple=None , ) ->List[Any]: '''simple docstring''' A__ = parent A__ = batch_size A__ = image_size A__ = num_channels A__ = embeddings_size A__ = hidden_sizes A__ = depths A__ = is_training A__ = use_labels A__ = hidden_act A__ = num_labels A__ = scope A__ = len(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Tuple: '''simple docstring''' A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.num_labels) A__ = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str) ->Any: '''simple docstring''' return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]) ->Tuple: '''simple docstring''' A__ = TFRegNetModel(config=UpperCAmelCase__) A__ = model(UpperCAmelCase__ , training=UpperCAmelCase__) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict) ->Any: '''simple docstring''' A__ = self.num_labels A__ = TFRegNetForImageClassification(UpperCAmelCase__) A__ = model(UpperCAmelCase__ , labels=UpperCAmelCase__ , training=UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def SCREAMING_SNAKE_CASE ( self : Dict) ->Optional[int]: '''simple docstring''' A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class UpperCamelCase_ ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () UpperCAmelCase__ = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def SCREAMING_SNAKE_CASE ( self : List[str]) ->Tuple: '''simple docstring''' A__ = TFRegNetModelTester(self) A__ = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[str]) ->Tuple: '''simple docstring''' return @unittest.skip(reason='''RegNet does not use inputs_embeds''') def SCREAMING_SNAKE_CASE ( self : Tuple) ->int: '''simple docstring''' pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('''GPU''')) == 0 , reason='''TF does not support backprop for grouped convolutions on CPU.''' , ) @slow def SCREAMING_SNAKE_CASE ( self : str) ->str: '''simple docstring''' super().test_keras_fit() @unittest.skip(reason='''RegNet does not support input and output embeddings''') def SCREAMING_SNAKE_CASE ( self : Dict) ->int: '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self : str) ->List[str]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(UpperCAmelCase__) A__ = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : int) ->Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Dict: '''simple docstring''' def check_hidden_states_output(UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[Any]): A__ = model_class(UpperCAmelCase__) A__ = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) , training=UpperCAmelCase__) A__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A__ = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase__) , expected_num_stages + 1) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: A__ = layer_type A__ = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Dict) ->int: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict={}): A__ = model(UpperCAmelCase__ , return_dict=UpperCAmelCase__ , **UpperCAmelCase__) A__ = model(UpperCAmelCase__ , return_dict=UpperCAmelCase__ , **UpperCAmelCase__).to_tuple() def recursive_check(UpperCAmelCase__ : int , UpperCAmelCase__ : List[str]): if isinstance(UpperCAmelCase__ , (List, Tuple)): for tuple_iterable_value, dict_iterable_value in zip(UpperCAmelCase__ , UpperCAmelCase__): recursive_check(UpperCAmelCase__ , UpperCAmelCase__) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(UpperCAmelCase__ , UpperCAmelCase__)) , msg=( '''Tuple and dict output are not equal. Difference:''' f""" {tf.math.reduce_max(tf.abs(tuple_object - dict_object))}""" ) , ) recursive_check(UpperCAmelCase__ , UpperCAmelCase__) for model_class in self.all_model_classes: A__ = model_class(UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) check_equivalence(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__) check_equivalence(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__) check_equivalence(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , {'''output_hidden_states''': True}) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__) A__ = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__) check_equivalence(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , {'''output_hidden_states''': True}) def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->int: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase__) @slow def SCREAMING_SNAKE_CASE ( self : List[Any]) ->List[str]: '''simple docstring''' for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = TFRegNetModel.from_pretrained(UpperCAmelCase__) self.assertIsNotNone(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: """simple docstring""" A__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE ( self : Dict) ->List[Any]: '''simple docstring''' return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->int: '''simple docstring''' A__ = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=UpperCAmelCase__ , return_tensors='''tf''') # forward pass A__ = model(**UpperCAmelCase__ , training=UpperCAmelCase__) # verify the logits A__ = tf.TensorShape((1, 1_000)) self.assertEqual(outputs.logits.shape , UpperCAmelCase__) A__ = tf.constant([-0.4180, -1.5051, -3.4836]) tf.debugging.assert_near(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4)
87
from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class UpperCamelCase_ ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase__ = ( { '''feature-extraction''': TFMobileBertModel, '''fill-mask''': TFMobileBertForMaskedLM, '''question-answering''': TFMobileBertForQuestionAnswering, '''text-classification''': TFMobileBertForSequenceClassification, '''token-classification''': TFMobileBertForTokenClassification, '''zero-shot''': TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : str=False) ->Optional[Any]: '''simple docstring''' A__ = super()._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__) if return_labels: if model_class in get_values(UpperCAmelCase__): A__ = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa) return inputs_dict class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : int=13 , UpperCAmelCase__ : str=7 , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : str=99 , UpperCAmelCase__ : List[str]=32 , UpperCAmelCase__ : Optional[int]=32 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : List[str]=4 , UpperCAmelCase__ : Optional[Any]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : List[Any]=512 , UpperCAmelCase__ : Tuple=16 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Dict=0.02 , UpperCAmelCase__ : int=3 , UpperCAmelCase__ : List[str]=4 , UpperCAmelCase__ : Tuple=None , ) ->Any: '''simple docstring''' A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = scope A__ = embedding_size def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Tuple: '''simple docstring''' A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length]) A__ = None if self.use_token_type_ids: A__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size) A__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) A__ = ids_tensor([self.batch_size] , self.num_choices) A__ = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any]) ->Any: '''simple docstring''' A__ = TFMobileBertModel(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) A__ = [input_ids, input_mask] A__ = model(UpperCAmelCase__) A__ = model(UpperCAmelCase__) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Tuple) ->Optional[Any]: '''simple docstring''' A__ = TFMobileBertForMaskedLM(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any]) ->int: '''simple docstring''' A__ = TFMobileBertForNextSentencePrediction(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2)) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int) ->List[Any]: '''simple docstring''' A__ = TFMobileBertForPreTraining(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2)) def SCREAMING_SNAKE_CASE ( self : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Tuple) ->Dict: '''simple docstring''' A__ = self.num_labels A__ = TFMobileBertForSequenceClassification(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def SCREAMING_SNAKE_CASE ( self : Any , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : int) ->Dict: '''simple docstring''' A__ = self.num_choices A__ = TFMobileBertForMultipleChoice(config=UpperCAmelCase__) A__ = tf.tile(tf.expand_dims(UpperCAmelCase__ , 1) , (1, self.num_choices, 1)) A__ = tf.tile(tf.expand_dims(UpperCAmelCase__ , 1) , (1, self.num_choices, 1)) A__ = tf.tile(tf.expand_dims(UpperCAmelCase__ , 1) , (1, self.num_choices, 1)) A__ = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int]) ->int: '''simple docstring''' A__ = self.num_labels A__ = TFMobileBertForTokenClassification(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def SCREAMING_SNAKE_CASE ( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]) ->Union[str, Any]: '''simple docstring''' A__ = TFMobileBertForQuestionAnswering(config=UpperCAmelCase__) A__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} A__ = model(UpperCAmelCase__) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def SCREAMING_SNAKE_CASE ( self : Any) ->str: '''simple docstring''' A__ = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) = config_and_inputs A__ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self : Any) ->Union[str, Any]: '''simple docstring''' A__ = TFMobileBertModelTest.TFMobileBertModelTester(self) A__ = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=37) def SCREAMING_SNAKE_CASE ( self : Tuple) ->Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Dict: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[str]) ->Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Dict: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Tuple) ->Dict: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[Any]) ->Union[str, Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : int) ->Optional[int]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Tuple) ->Optional[int]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Tuple) ->List[str]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*UpperCAmelCase__) @slow def SCREAMING_SNAKE_CASE ( self : str) ->List[Any]: '''simple docstring''' for model_name in ["google/mobilebert-uncased"]: A__ = TFMobileBertModel.from_pretrained(UpperCAmelCase__) self.assertIsNotNone(UpperCAmelCase__) @require_tf class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Any: '''simple docstring''' A__ = TFMobileBertForPreTraining.from_pretrained('''google/mobilebert-uncased''') A__ = tf.constant([[0, 1, 2, 3, 4, 5]]) A__ = model(UpperCAmelCase__)[0] A__ = [1, 6, 30_522] self.assertEqual(output.shape , UpperCAmelCase__) A__ = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ]) tf.debugging.assert_near(output[:, :3, :3] , UpperCAmelCase__ , atol=1e-4)
87
1
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class snake_case_ ( unittest.TestCase ): def __A ( self , __lowerCAmelCase ): for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] , model_result['ss'] ): SCREAMING_SNAKE_CASE_ : str = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(__lowerCAmelCase ) def __A ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : List[str] = 'sgugger/tiny-distilbert-classification' SCREAMING_SNAKE_CASE_ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , only_pretrain_model=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Tuple = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : List[str] = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , torchscript=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : List[Any] = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def __A ( self ): SCREAMING_SNAKE_CASE_ : Any = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , fpaa=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : int = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : List[Any] = AutoConfig.from_pretrained(__lowerCAmelCase ) # set architectures equal to `None` SCREAMING_SNAKE_CASE_ : Any = None SCREAMING_SNAKE_CASE_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Optional[Any] = PyTorchBenchmark(__lowerCAmelCase , configs=[config] ) SCREAMING_SNAKE_CASE_ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : int = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Any = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == 'cpu' , 'Can\'t do half precision' ) def __A ( self ): SCREAMING_SNAKE_CASE_ : str = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=__lowerCAmelCase , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Tuple = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : Any = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : Tuple = AutoConfig.from_pretrained(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : List[Any] = PyTorchBenchmark(__lowerCAmelCase , configs=[config] ) SCREAMING_SNAKE_CASE_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : List[str] = 'sshleifer/tinier_bart' SCREAMING_SNAKE_CASE_ : Union[str, Any] = AutoConfig.from_pretrained(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : List[str] = PyTorchBenchmark(__lowerCAmelCase , configs=[config] ) SCREAMING_SNAKE_CASE_ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : Tuple = 'sshleifer/tiny-gpt2' SCREAMING_SNAKE_CASE_ : List[Any] = AutoConfig.from_pretrained(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = PyTorchBenchmark(__lowerCAmelCase , configs=[config] ) SCREAMING_SNAKE_CASE_ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : Tuple = 'sshleifer/tinier_bart' SCREAMING_SNAKE_CASE_ : Dict = AutoConfig.from_pretrained(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmark(__lowerCAmelCase , configs=[config] ) SCREAMING_SNAKE_CASE_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): SCREAMING_SNAKE_CASE_ : Dict = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , save_to_csv=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(__lowerCAmelCase , 'inf_time.csv' ) , train_memory_csv_file=os.path.join(__lowerCAmelCase , 'train_mem.csv' ) , inference_memory_csv_file=os.path.join(__lowerCAmelCase , 'inf_mem.csv' ) , train_time_csv_file=os.path.join(__lowerCAmelCase , 'train_time.csv' ) , env_info_csv_file=os.path.join(__lowerCAmelCase , 'env.csv' ) , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Any = PyTorchBenchmark(__lowerCAmelCase ) benchmark.run() self.assertTrue(Path(os.path.join(__lowerCAmelCase , 'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__lowerCAmelCase , 'train_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__lowerCAmelCase , 'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__lowerCAmelCase , 'train_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__lowerCAmelCase , 'env.csv' ) ).exists() ) def __A ( self ): SCREAMING_SNAKE_CASE_ : int = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(__lowerCAmelCase ): self.assertTrue(hasattr(__lowerCAmelCase , 'sequential' ) ) self.assertTrue(hasattr(__lowerCAmelCase , 'cumulative' ) ) self.assertTrue(hasattr(__lowerCAmelCase , 'current' ) ) self.assertTrue(hasattr(__lowerCAmelCase , 'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=__lowerCAmelCase , inference=__lowerCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(__lowerCAmelCase , 'log.txt' ) , log_print=__lowerCAmelCase , trace_memory_line_by_line=__lowerCAmelCase , multi_process=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = PyTorchBenchmark(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Dict = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(__lowerCAmelCase , 'log.txt' ) ).exists() )
715
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase__: Optional[Any] = {"configuration_plbart": ["PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "PLBartConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__: str = ["PLBartTokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__: Any = [ "PLBART_PRETRAINED_MODEL_ARCHIVE_LIST", "PLBartForCausalLM", "PLBartForConditionalGeneration", "PLBartForSequenceClassification", "PLBartModel", "PLBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys lowerCAmelCase__: Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure)
311
0
"""simple docstring""" import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration _lowerCAmelCase : int = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def lowerCamelCase_( _lowerCamelCase ) -> Dict: '''simple docstring''' _lowerCamelCase : Tuple = ["layers", "blocks"] for k in ignore_keys: state_dict.pop(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : int = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def lowerCamelCase_( _lowerCamelCase ) -> Any: '''simple docstring''' _lowerCamelCase : Union[str, Any] = list(s_dict.keys() ) for key in keys: _lowerCamelCase : int = key for k, v in WHISPER_MAPPING.items(): if k in key: _lowerCamelCase : List[Any] = new_key.replace(_lowerCamelCase , _lowerCamelCase ) print(F"""{key} -> {new_key}""" ) _lowerCamelCase : Union[str, Any] = s_dict.pop(_lowerCamelCase ) return s_dict def lowerCamelCase_( _lowerCamelCase ) -> Any: '''simple docstring''' _lowerCamelCase, _lowerCamelCase : Any = emb.weight.shape _lowerCamelCase : Union[str, Any] = nn.Linear(_lowerCamelCase , _lowerCamelCase , bias=_lowerCamelCase ) _lowerCamelCase : List[Any] = emb.weight.data return lin_layer def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> bytes: '''simple docstring''' os.makedirs(_lowerCamelCase , exist_ok=_lowerCamelCase ) _lowerCamelCase : Optional[Any] = os.path.basename(_lowerCamelCase ) _lowerCamelCase : Union[str, Any] = url.split("/" )[-2] _lowerCamelCase : Union[str, Any] = os.path.join(_lowerCamelCase , _lowerCamelCase ) if os.path.exists(_lowerCamelCase ) and not os.path.isfile(_lowerCamelCase ): raise RuntimeError(F"""{download_target} exists and is not a regular file""" ) if os.path.isfile(_lowerCamelCase ): _lowerCamelCase : List[Any] = open(_lowerCamelCase , "rb" ).read() if hashlib.shaaaa(_lowerCamelCase ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(F"""{download_target} exists, but the SHA256 checksum does not match; re-downloading the file""" ) with urllib.request.urlopen(_lowerCamelCase ) as source, open(_lowerCamelCase , "wb" ) as output: with tqdm( total=int(source.info().get("Content-Length" ) ) , ncols=80 , unit="iB" , unit_scale=_lowerCamelCase , unit_divisor=1024 ) as loop: while True: _lowerCamelCase : str = source.read(8192 ) if not buffer: break output.write(_lowerCamelCase ) loop.update(len(_lowerCamelCase ) ) _lowerCamelCase : Dict = open(_lowerCamelCase , "rb" ).read() if hashlib.shaaaa(_lowerCamelCase ).hexdigest() != expected_shaaaa: raise RuntimeError( "Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model." ) return model_bytes def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> str: '''simple docstring''' if ".pt" not in checkpoint_path: _lowerCamelCase : Tuple = _download(_MODELS[checkpoint_path] ) else: _lowerCamelCase : List[Any] = torch.load(_lowerCamelCase , map_location="cpu" ) _lowerCamelCase : Dict = original_checkpoint["dims"] _lowerCamelCase : Union[str, Any] = original_checkpoint["model_state_dict"] _lowerCamelCase : int = state_dict["decoder.token_embedding.weight"] remove_ignore_keys_(_lowerCamelCase ) rename_keys(_lowerCamelCase ) _lowerCamelCase : Union[str, Any] = True _lowerCamelCase : List[str] = state_dict["decoder.layers.0.fc1.weight"].shape[0] _lowerCamelCase : Dict = WhisperConfig( vocab_size=dimensions["n_vocab"] , encoder_ffn_dim=_lowerCamelCase , decoder_ffn_dim=_lowerCamelCase , num_mel_bins=dimensions["n_mels"] , d_model=dimensions["n_audio_state"] , max_target_positions=dimensions["n_text_ctx"] , encoder_layers=dimensions["n_audio_layer"] , encoder_attention_heads=dimensions["n_audio_head"] , decoder_layers=dimensions["n_text_layer"] , decoder_attention_heads=dimensions["n_text_state"] , max_source_positions=dimensions["n_audio_ctx"] , ) _lowerCamelCase : List[str] = WhisperForConditionalGeneration(_lowerCamelCase ) _lowerCamelCase, _lowerCamelCase : int = model.model.load_state_dict(_lowerCamelCase , strict=_lowerCamelCase ) if len(_lowerCamelCase ) > 0 and not set(_lowerCamelCase ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( "Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing," F""" but all the following weights are missing {missing}""" ) if tie_embeds: _lowerCamelCase : Optional[int] = make_linear_from_emb(model.model.decoder.embed_tokens ) else: _lowerCamelCase : Tuple = proj_out_weights model.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": _lowerCAmelCase : Dict = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _lowerCAmelCase : int = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
46
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[Any] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'huggingface/autoformer-tourism-monthly': 'https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json', } class lowercase_ (lowercase__ ): snake_case ='autoformer' snake_case ={ 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self , lowercase_ = None , lowercase_ = None , lowercase_ = "student_t" , lowercase_ = "nll" , lowercase_ = 1 , lowercase_ = [1, 2, 3, 4, 5, 6, 7] , lowercase_ = True , lowercase_ = 0 , lowercase_ = 0 , lowercase_ = 0 , lowercase_ = 0 , lowercase_ = None , lowercase_ = None , lowercase_ = 64 , lowercase_ = 2 , lowercase_ = 2 , lowercase_ = 2 , lowercase_ = 2 , lowercase_ = 32 , lowercase_ = 32 , lowercase_ = "gelu" , lowercase_ = 0.1 , lowercase_ = 0.1 , lowercase_ = 0.1 , lowercase_ = 0.1 , lowercase_ = 0.1 , lowercase_ = 100 , lowercase_ = 0.02 , lowercase_ = True , lowercase_=True , lowercase_ = 10 , lowercase_ = 25 , lowercase_ = 3 , **lowercase_ , ) -> Union[str, Any]: # time series specific configuration a__ =prediction_length a__ =context_length if context_length is not None else prediction_length a__ =distribution_output a__ =loss a__ =input_size a__ =num_time_features a__ =lags_sequence a__ =scaling a__ =num_dynamic_real_features a__ =num_static_real_features a__ =num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(lowercase_) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`') a__ =cardinality else: a__ =[0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(lowercase_) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`') a__ =embedding_dimension else: a__ =[min(50 , (cat + 1) // 2) for cat in self.cardinality] a__ =num_parallel_samples # Transformer architecture configuration a__ =input_size * len(self.lags_sequence) + self._number_of_features a__ =d_model a__ =encoder_attention_heads a__ =decoder_attention_heads a__ =encoder_ffn_dim a__ =decoder_ffn_dim a__ =encoder_layers a__ =decoder_layers a__ =dropout a__ =attention_dropout a__ =activation_dropout a__ =encoder_layerdrop a__ =decoder_layerdrop a__ =activation_function a__ =init_std a__ =use_cache # Autoformer a__ =label_length a__ =moving_average a__ =autocorrelation_factor super().__init__(is_encoder_decoder=lowercase_ , **lowercase_) @property def __UpperCamelCase ( self) -> int: return ( sum(self.embedding_dimension) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
20
0
"""simple docstring""" import argparse import math import traceback import dateutil.parser as date_parser import requests def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int] ) ->Tuple: A__ : Optional[int] = {} A__ : int = job["""started_at"""] A__ : Any = job["""completed_at"""] A__ : Optional[int] = date_parser.parse(UpperCAmelCase__ ) A__ : Optional[Any] = date_parser.parse(UpperCAmelCase__ ) A__ : Any = round((end_datetime - start_datetime).total_seconds() / 60.0 ) A__ : List[str] = start A__ : Optional[int] = end A__ : Optional[int] = duration_in_min return job_info def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Optional[int]=None ) ->Optional[int]: A__ : List[Any] = None if token is not None: A__ : Optional[Any] = {"""Accept""": """application/vnd.github+json""", """Authorization""": f'Bearer {token}'} A__ : Any = f'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' A__ : Any = requests.get(UpperCAmelCase__, headers=UpperCAmelCase__ ).json() A__ : int = {} try: job_time.update({job["""name"""]: extract_time_from_single_job(UpperCAmelCase__ ) for job in result["""jobs"""]} ) A__ : str = math.ceil((result["""total_count"""] - 1_0_0) / 1_0_0 ) for i in range(UpperCAmelCase__ ): A__ : Dict = requests.get(url + f'&page={i + 2}', headers=UpperCAmelCase__ ).json() job_time.update({job["""name"""]: extract_time_from_single_job(UpperCAmelCase__ ) for job in result["""jobs"""]} ) return job_time except Exception: print(f'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') A_ = parser.parse_args() A_ = get_job_time(args.workflow_run_id) A_ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(F'{k}: {v["duration"]}')
498
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->str: if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) A__ : str = str(bin(UpperCAmelCase__ ) )[2:] # remove the leading "0b" A__ : List[Any] = str(bin(UpperCAmelCase__ ) )[2:] # remove the leading "0b" A__ : List[str] = max(len(UpperCAmelCase__ ), len(UpperCAmelCase__ ) ) return "0b" + "".join( str(int(char_a == """1""" and char_b == """1""" ) ) for char_a, char_b in zip(a_binary.zfill(UpperCAmelCase__ ), b_binary.zfill(UpperCAmelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
498
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class _UpperCamelCase ( __snake_case): def __init__(self , lowerCamelCase__ ): """simple docstring""" A__ = data def __iter__(self ): """simple docstring""" for element in self.data: yield element def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Dict=True ): A__ = Accelerator(even_batches=UpperCamelCase ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Accelerator , UpperCamelCase : int , UpperCamelCase : int , UpperCamelCase : bool = False ): if iterable: A__ = DummyIterableDataset(torch.as_tensor(range(UpperCamelCase ) ) ) else: A__ = TensorDataset(torch.as_tensor(range(UpperCamelCase ) ) ) A__ = DataLoader(UpperCamelCase , batch_size=UpperCamelCase ) A__ = accelerator.prepare(UpperCamelCase ) return dl def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Accelerator , UpperCamelCase : int , UpperCamelCase : int , UpperCamelCase : List[int] , UpperCamelCase : List[int] , ): A__ = create_dataloader(accelerator=UpperCamelCase , dataset_size=UpperCamelCase , batch_size=UpperCamelCase ) A__ = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def _SCREAMING_SNAKE_CASE ( ): A__ = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( UpperCamelCase , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( UpperCamelCase , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def _SCREAMING_SNAKE_CASE ( ): A__ = create_accelerator(even_batches=UpperCamelCase ) verify_dataloader_batch_sizes( UpperCamelCase , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( UpperCamelCase , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def _SCREAMING_SNAKE_CASE ( ): A__ = create_accelerator(even_batches=UpperCamelCase ) A__ = torch.nn.Linear(1 , 1 ) A__ = accelerator.prepare(UpperCamelCase ) A__ = create_dataloader(UpperCamelCase , dataset_size=3 , batch_size=1 ) A__ = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(UpperCamelCase ): A__ = ddp_model(batch[0].float() ) A__ = output.sum() loss.backward() batch_idxs.append(UpperCamelCase ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Dict ): with warnings.catch_warnings(record=UpperCamelCase ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , UpperCamelCase ) assert "only supported for multi-GPU" in str(w[-1].message ) def _SCREAMING_SNAKE_CASE ( ): A__ = True A__ = False A__ = create_accelerator(even_batches=UpperCamelCase ) A__ = torch.nn.Linear(1 , 1 ) A__ = accelerator.prepare(UpperCamelCase ) A__ = create_dataloader(UpperCamelCase , dataset_size=3 , batch_size=1 ) A__ = create_dataloader(UpperCamelCase , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=UpperCamelCase ): A__ = train_dl.batch_sampler.even_batches A__ = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def _SCREAMING_SNAKE_CASE ( ): A__ = True A__ = False A__ = create_accelerator(even_batches=UpperCamelCase ) A__ = torch.nn.Linear(1 , 1 ) A__ = accelerator.prepare(UpperCamelCase ) create_dataloader(UpperCamelCase , dataset_size=3 , batch_size=1 , iterable=UpperCamelCase ) A__ = create_dataloader(UpperCamelCase , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings("""ignore""" ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=UpperCamelCase ): A__ = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def _SCREAMING_SNAKE_CASE ( ): A__ = create_accelerator() A__ = torch.nn.Linear(1 , 1 ) A__ = accelerator.prepare(UpperCamelCase ) create_dataloader(UpperCamelCase , dataset_size=3 , batch_size=1 , iterable=UpperCamelCase ) with warnings.catch_warnings(record=UpperCamelCase ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=UpperCamelCase ): pass assert issubclass(w[-1].category , UpperCamelCase ) assert "only supported for map-style datasets" in str(w[-1].message ) def _SCREAMING_SNAKE_CASE ( ): A__ = create_accelerator() accelerator.print("""Test that even_batches variable ensures uniform batches across processes""" ) test_default_ensures_even_batch_sizes() accelerator.print("""Run tests with even_batches disabled""" ) test_can_disable_even_batches() accelerator.print("""Test joining uneven inputs""" ) test_can_join_uneven_inputs() accelerator.print("""Test overriding even_batches when joining uneven inputs""" ) test_join_can_override_even_batches() accelerator.print("""Test overriding even_batches for mixed dataloader types""" ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print("""Test overriding even_batches raises a warning for iterable dataloaders""" ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print("""Test join with non DDP distributed raises warning""" ) A__ = accelerator.state.distributed_type A__ = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(UpperCamelCase ) A__ = original_state if __name__ == "__main__": main()
574
"""simple docstring""" import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): lowerCamelCase__ = yaml.safe_load( "\\nname: \"\"\nallow_empty: false\nallow_empty_text: true\nsubsections:\n - name: \"Dataset Card for X\" # First-level markdown heading\n allow_empty: false\n allow_empty_text: true\n subsections:\n - name: \"Table of Contents\"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: \"Dataset Description\"\n allow_empty: false\n allow_empty_text: false\n subsections:\n - name: \"Dataset Summary\"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: \"Supported Tasks and Leaderboards\"\n allow_empty: true\n allow_empty_text: true\n subsections: null\n - name: Languages\n allow_empty: false\n allow_empty_text: true\n subsections: null\n" ) lowerCamelCase__ = { "name": "root", "text": "", "is_empty_text": True, "subsections": [ { "name": "Dataset Card for My Dataset", "text": "", "is_empty_text": True, "subsections": [ {"name": "Table of Contents", "text": "Some text here.", "is_empty_text": False, "subsections": []}, { "name": "Dataset Description", "text": "Some text here.", "is_empty_text": False, "subsections": [ { "name": "Dataset Summary", "text": "Some text here.", "is_empty_text": False, "subsections": [], }, { "name": "Supported Tasks and Leaderboards", "text": "", "is_empty_text": True, "subsections": [], }, {"name": "Languages", "text": "Language Text", "is_empty_text": False, "subsections": []}, ], }, ], } ], } lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n#### Extra Ignored Subsection\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = { "name": "root", "text": "", "is_empty_text": True, "subsections": [ { "name": "Dataset Card for My Dataset", "text": "", "is_empty_text": True, "subsections": [ {"name": "Table of Contents", "text": "Some text here.", "is_empty_text": False, "subsections": []}, { "name": "Dataset Description", "text": "Some text here.", "is_empty_text": False, "subsections": [ { "name": "Dataset Summary", "text": "Some text here.", "is_empty_text": False, "subsections": [ { "name": "Extra Ignored Subsection", "text": "", "is_empty_text": True, "subsections": [], } ], }, { "name": "Supported Tasks and Leaderboards", "text": "", "is_empty_text": True, "subsections": [], }, {"name": "Languages", "text": "Language Text", "is_empty_text": False, "subsections": []}, ], }, ], } ], } lowerCamelCase__ = "\\n---\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = ( "The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README." ) lowerCamelCase__ = "\\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = ( "The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README." ) lowerCamelCase__ = "\\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored)." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found 'None'." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Languages\nLanguage Text\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n# Dataset Card My Dataset\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README." lowerCamelCase__ = "" lowerCamelCase__ = "The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README." lowerCamelCase__ = "\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n" lowerCamelCase__ = "The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections." @pytest.mark.parametrize( """readme_md, expected_dict""" , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Any , UpperCamelCase : int ): assert ReadMe.from_string(UpperCamelCase , UpperCamelCase ).to_dict() == expected_dict @pytest.mark.parametrize( """readme_md, expected_error""" , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : int , UpperCamelCase : Optional[Any] ): with pytest.raises(UpperCamelCase , match=re.escape(expected_error.format(path="""root""" ) ) ): A__ = ReadMe.from_string(UpperCamelCase , UpperCamelCase ) readme.validate() @pytest.mark.parametrize( """readme_md, expected_error""" , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : List[str] , UpperCamelCase : str ): with pytest.raises(UpperCamelCase , match=re.escape(expected_error.format(path="""root""" ) ) ): ReadMe.from_string(UpperCamelCase , UpperCamelCase ) @pytest.mark.parametrize( """readme_md,""" , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : str ): ReadMe.from_string(UpperCamelCase , UpperCamelCase , suppress_parsing_errors=UpperCamelCase ) @pytest.mark.parametrize( """readme_md, expected_dict""" , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : List[str] , UpperCamelCase : Dict ): with tempfile.TemporaryDirectory() as tmp_dir: A__ = Path(UpperCamelCase ) / """README.md""" with open(UpperCamelCase , """w+""" ) as readme_file: readme_file.write(UpperCamelCase ) A__ = ReadMe.from_readme(UpperCamelCase , UpperCamelCase ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( """readme_md, expected_error""" , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Any , UpperCamelCase : Optional[int] ): with tempfile.TemporaryDirectory() as tmp_dir: A__ = Path(UpperCamelCase ) / """README.md""" with open(UpperCamelCase , """w+""" ) as readme_file: readme_file.write(UpperCamelCase ) A__ = expected_error.format(path=UpperCamelCase ) with pytest.raises(UpperCamelCase , match=re.escape(UpperCamelCase ) ): A__ = ReadMe.from_readme(UpperCamelCase , UpperCamelCase ) readme.validate() @pytest.mark.parametrize( """readme_md, expected_error""" , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Dict , UpperCamelCase : List[Any] ): with tempfile.TemporaryDirectory() as tmp_dir: A__ = Path(UpperCamelCase ) / """README.md""" with open(UpperCamelCase , """w+""" ) as readme_file: readme_file.write(UpperCamelCase ) A__ = expected_error.format(path=UpperCamelCase ) with pytest.raises(UpperCamelCase , match=re.escape(UpperCamelCase ) ): ReadMe.from_readme(UpperCamelCase , UpperCamelCase ) @pytest.mark.parametrize( """readme_md,""" , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Tuple ): with tempfile.TemporaryDirectory() as tmp_dir: A__ = Path(UpperCamelCase ) / """README.md""" with open(UpperCamelCase , """w+""" ) as readme_file: readme_file.write(UpperCamelCase ) ReadMe.from_readme(UpperCamelCase , UpperCamelCase , suppress_parsing_errors=UpperCamelCase )
574
1
"""simple docstring""" import argparse import struct import unittest class _lowercase : def __init__( self : str , a : bytes ): """simple docstring""" __snake_case : List[Any] =data # Initialize hash values __snake_case : Optional[Any] =[ 0x6A09E667, 0xBB67AE85, 0x3C6EF372, 0xA54FF53A, 0x510E527F, 0x9B05688C, 0x1F83D9AB, 0x5BE0CD19, ] # Initialize round constants __snake_case : List[str] =[ 0x428A2F98, 0x71374491, 0xB5C0FBCF, 0xE9B5DBA5, 0x3956C25B, 0x59F111F1, 0x923F82A4, 0xAB1C5ED5, 0xD807AA98, 0x12835B01, 0x243185BE, 0x550C7DC3, 0x72BE5D74, 0x80DEB1FE, 0x9BDC06A7, 0xC19BF174, 0xE49B69C1, 0xEFBE4786, 0x0FC19DC6, 0x240CA1CC, 0x2DE92C6F, 0x4A7484AA, 0x5CB0A9DC, 0x76F988DA, 0x983E5152, 0xA831C66D, 0xB00327C8, 0xBF597FC7, 0xC6E00BF3, 0xD5A79147, 0x06CA6351, 0x14292967, 0x27B70A85, 0x2E1B2138, 0x4D2C6DFC, 0x53380D13, 0x650A7354, 0x766A0ABB, 0x81C2C92E, 0x92722C85, 0xA2BFE8A1, 0xA81A664B, 0xC24B8B70, 0xC76C51A3, 0xD192E819, 0xD6990624, 0xF40E3585, 0x106AA070, 0x19A4C116, 0x1E376C08, 0x2748774C, 0x34B0BCB5, 0x391C0CB3, 0x4ED8AA4A, 0x5B9CCA4F, 0x682E6FF3, 0x748F82EE, 0x78A5636F, 0x84C87814, 0x8CC70208, 0x90BEFFFA, 0xA4506CEB, 0xBEF9A3F7, 0xC67178F2, ] __snake_case : List[Any] =self.preprocessing(self.data ) self.final_hash() @staticmethod def _UpperCamelCase ( a : bytes ): """simple docstring""" __snake_case : int =b'''\x80''' + (b'''\x00''' * (6_3 - (len(a ) + 8) % 6_4)) __snake_case : str =struct.pack('''>Q''' , (len(a ) * 8) ) return data + padding + big_endian_integer def _UpperCamelCase ( self : List[str] ): """simple docstring""" __snake_case : Tuple =[ self.preprocessed_data[x : x + 6_4] for x in range(0 , len(self.preprocessed_data ) , 6_4 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers __snake_case : Dict =list(struct.unpack('''>16L''' , a ) ) # add 48 0-ed integers words += [0] * 4_8 __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case : Optional[Any] =self.hashes for index in range(0 , 6_4 ): if index > 1_5: # modify the zero-ed indexes at the end of the array __snake_case : Union[str, Any] =( self.ror(words[index - 1_5] , 7 ) ^ self.ror(words[index - 1_5] , 1_8 ) ^ (words[index - 1_5] >> 3) ) __snake_case : Any =( self.ror(words[index - 2] , 1_7 ) ^ self.ror(words[index - 2] , 1_9 ) ^ (words[index - 2] >> 1_0) ) __snake_case : Any =( words[index - 1_6] + sa + words[index - 7] + sa ) % 0x100000000 # Compression __snake_case : Dict =self.ror(a , 6 ) ^ self.ror(a , 1_1 ) ^ self.ror(a , 2_5 ) __snake_case : Any =(e & f) ^ ((~e & 0xFFFFFFFF) & g) __snake_case : Dict =( h + sa + ch + self.round_constants[index] + words[index] ) % 0x100000000 __snake_case : Any =self.ror(a , 2 ) ^ self.ror(a , 1_3 ) ^ self.ror(a , 2_2 ) __snake_case : List[str] =(a & b) ^ (a & c) ^ (b & c) __snake_case : str =(sa + maj) % 0x100000000 __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case : Any =( g, f, e, ((d + tempa) % 0x100000000), c, b, a, ((tempa + tempa) % 0x100000000), ) __snake_case : Optional[Any] =[a, b, c, d, e, f, g, h] # Modify final values __snake_case : Any =[ ((element + mutated_hash_values[index]) % 0x100000000) for index, element in enumerate(self.hashes ) ] __snake_case : Optional[int] =''''''.join([hex(a )[2:].zfill(8 ) for value in self.hashes] ) def _UpperCamelCase ( self : Union[str, Any] , a : int , a : int ): """simple docstring""" return 0xFFFFFFFF & (value << (3_2 - rotations)) | (value >> rotations) class _lowercase ( unittest.TestCase ): def _UpperCamelCase ( self : str ): """simple docstring""" import hashlib __snake_case : Dict =bytes('''Test String''' , '''utf-8''' ) self.assertEqual(SHAaaa(a ).hash , hashlib.shaaaa(a ).hexdigest() ) def __lowercase ( ) -> None: import doctest doctest.testmod() __snake_case : Dict =argparse.ArgumentParser() parser.add_argument( '''-s''' , '''--string''' , dest='''input_string''' , default='''Hello World!! Welcome to Cryptography''' , help='''Hash the string''' , ) parser.add_argument( '''-f''' , '''--file''' , dest='''input_file''' , help='''Hash contents of a file''' ) __snake_case : str =parser.parse_args() __snake_case : Tuple =args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , '''rb''' ) as f: __snake_case : Union[str, Any] =f.read() else: __snake_case : int =bytes(a , '''utf-8''' ) print(SHAaaa(a ).hash ) if __name__ == "__main__": main()
497
"""simple docstring""" from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance UpperCamelCase_ : Optional[Any] = 6_37_81_37.0 UpperCamelCase_ : Optional[Any] = 6_35_67_52.31_42_45 UpperCamelCase_ : Union[str, Any] = 6378137 def __lowercase ( a : float , a : float , a : float , a : float ) -> float: __snake_case : Optional[Any] =(AXIS_A - AXIS_B) / AXIS_A # Parametric latitudes # https://en.wikipedia.org/wiki/Latitude#Parametric_(or_reduced)_latitude __snake_case : List[str] =atan((1 - flattening) * tan(radians(a ) ) ) __snake_case : Any =atan((1 - flattening) * tan(radians(a ) ) ) # Compute central angle between two points # using haversine theta. sigma = haversine_distance / equatorial radius __snake_case : Optional[Any] =haversine_distance(a , a , a , a ) / EQUATORIAL_RADIUS # Intermediate P and Q values __snake_case : List[str] =(b_lata + b_lata) / 2 __snake_case : Dict =(b_lata - b_lata) / 2 # Intermediate X value # X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2) __snake_case : Optional[int] =(sin(a ) ** 2) * (cos(a ) ** 2) __snake_case : Any =cos(sigma / 2 ) ** 2 __snake_case : Optional[Any] =(sigma - sin(a )) * (x_numerator / x_demonimator) # Intermediate Y value # Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2) __snake_case : Tuple =(cos(a ) ** 2) * (sin(a ) ** 2) __snake_case : Dict =sin(sigma / 2 ) ** 2 __snake_case : Tuple =(sigma + sin(a )) * (y_numerator / y_denominator) return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value))) if __name__ == "__main__": import doctest doctest.testmod()
497
1
"""simple docstring""" def a ( __snake_case : int, __snake_case : int, __snake_case : int ): '''simple docstring''' UpperCAmelCase_ :Tuple = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def a ( ): '''simple docstring''' print(sum_of_series(1, 1, 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
608
import gc import inspect import unittest import torch from parameterized import parameterized from diffusers import PriorTransformer from diffusers.utils import floats_tensor, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin enable_full_determinism() class lowercase_ ( A , unittest.TestCase ): __lowerCamelCase = PriorTransformer __lowerCamelCase = "hidden_states" @property def _snake_case ( self ) -> List[str]: SCREAMING_SNAKE_CASE_ : List[str] =4 SCREAMING_SNAKE_CASE_ : Optional[int] =8 SCREAMING_SNAKE_CASE_ : Optional[Any] =7 SCREAMING_SNAKE_CASE_ : Dict =floats_tensor((batch_size, embedding_dim) ).to(__A ) SCREAMING_SNAKE_CASE_ : Dict =floats_tensor((batch_size, embedding_dim) ).to(__A ) SCREAMING_SNAKE_CASE_ : Dict =floats_tensor((batch_size, num_embeddings, embedding_dim) ).to(__A ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def _snake_case ( self , __A=0 ) -> int: torch.manual_seed(__A ) SCREAMING_SNAKE_CASE_ : str =4 SCREAMING_SNAKE_CASE_ : Union[str, Any] =8 SCREAMING_SNAKE_CASE_ : List[Any] =7 SCREAMING_SNAKE_CASE_ : Tuple =torch.randn((batch_size, embedding_dim) ).to(__A ) SCREAMING_SNAKE_CASE_ : int =torch.randn((batch_size, embedding_dim) ).to(__A ) SCREAMING_SNAKE_CASE_ : List[Any] =torch.randn((batch_size, num_embeddings, embedding_dim) ).to(__A ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } @property def _snake_case ( self ) -> Union[str, Any]: return (4, 8) @property def _snake_case ( self ) -> int: return (4, 8) def _snake_case ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ : List[Any] ={ '''num_attention_heads''': 2, '''attention_head_dim''': 4, '''num_layers''': 2, '''embedding_dim''': 8, '''num_embeddings''': 7, '''additional_embeddings''': 4, } SCREAMING_SNAKE_CASE_ : Union[str, Any] =self.dummy_input return init_dict, inputs_dict def _snake_case ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Any =PriorTransformer.from_pretrained( '''hf-internal-testing/prior-dummy''' , output_loading_info=__A ) self.assertIsNotNone(__A ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(__A ) SCREAMING_SNAKE_CASE_ : Optional[int] =model(**self.dummy_input )[0] assert hidden_states is not None, "Make sure output is not None" def _snake_case ( self ) -> str: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[Any] =self.prepare_init_args_and_inputs_for_common() SCREAMING_SNAKE_CASE_ : Optional[int] =self.model_class(**__A ) SCREAMING_SNAKE_CASE_ : List[Any] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ : Union[str, Any] =[*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ : Optional[int] =['''hidden_states''', '''timestep'''] self.assertListEqual(arg_names[:2] , __A ) def _snake_case ( self ) -> Dict: SCREAMING_SNAKE_CASE_ : Dict =PriorTransformer.from_pretrained('''hf-internal-testing/prior-dummy''' ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =model.to(__A ) if hasattr(__A , '''set_default_attn_processor''' ): model.set_default_attn_processor() SCREAMING_SNAKE_CASE_ : List[Any] =self.get_dummy_seed_input() with torch.no_grad(): SCREAMING_SNAKE_CASE_ : str =model(**__A )[0] SCREAMING_SNAKE_CASE_ : Any =output[0, :5].flatten().cpu() print(__A ) # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. SCREAMING_SNAKE_CASE_ : int =torch.tensor([-1.3_436, -0.2_870, 0.7_538, 0.4_368, -0.0_239] ) self.assertTrue(torch_all_close(__A , __A , rtol=1e-2 ) ) @slow class lowercase_ ( unittest.TestCase ): def _snake_case ( self , __A=1 , __A=768 , __A=77 , __A=0 ) -> str: torch.manual_seed(__A ) SCREAMING_SNAKE_CASE_ : Dict =batch_size SCREAMING_SNAKE_CASE_ : List[str] =embedding_dim SCREAMING_SNAKE_CASE_ : Optional[int] =num_embeddings SCREAMING_SNAKE_CASE_ : Optional[Any] =torch.randn((batch_size, embedding_dim) ).to(__A ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =torch.randn((batch_size, embedding_dim) ).to(__A ) SCREAMING_SNAKE_CASE_ : List[str] =torch.randn((batch_size, num_embeddings, embedding_dim) ).to(__A ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def _snake_case ( self ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @parameterized.expand( [ # fmt: off [13, [-0.5_861, 0.1_283, -0.0_931, 0.0_882, 0.4_476, 0.1_329, -0.0_498, 0.0_640]], [37, [-0.4_913, 0.0_110, -0.0_483, 0.0_541, 0.4_954, -0.0_170, 0.0_354, 0.1_651]], # fmt: on ] ) def _snake_case ( self , __A , __A ) -> Optional[int]: SCREAMING_SNAKE_CASE_ : Dict =PriorTransformer.from_pretrained('''kandinsky-community/kandinsky-2-1-prior''' , subfolder='''prior''' ) model.to(__A ) SCREAMING_SNAKE_CASE_ : Optional[Any] =self.get_dummy_seed_input(seed=__A ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ : Dict =model(**__A )[0] assert list(sample.shape ) == [1, 768] SCREAMING_SNAKE_CASE_ : Dict =sample[0, :8].flatten().cpu() print(__A ) SCREAMING_SNAKE_CASE_ : Optional[Any] =torch.tensor(__A ) assert torch_all_close(__A , __A , atol=1e-3 )
443
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Tuple = { "configuration_albert": ["ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "AlbertConfig", "AlbertOnnxConfig"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Optional[Any] = ["AlbertTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[Any] = ["AlbertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[Any] = [ "ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "AlbertForMaskedLM", "AlbertForMultipleChoice", "AlbertForPreTraining", "AlbertForQuestionAnswering", "AlbertForSequenceClassification", "AlbertForTokenClassification", "AlbertModel", "AlbertPreTrainedModel", "load_tf_weights_in_albert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Optional[Any] = [ "TF_ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFAlbertForMaskedLM", "TFAlbertForMultipleChoice", "TFAlbertForPreTraining", "TFAlbertForQuestionAnswering", "TFAlbertForSequenceClassification", "TFAlbertForTokenClassification", "TFAlbertMainLayer", "TFAlbertModel", "TFAlbertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = [ "FlaxAlbertForMaskedLM", "FlaxAlbertForMultipleChoice", "FlaxAlbertForPreTraining", "FlaxAlbertForQuestionAnswering", "FlaxAlbertForSequenceClassification", "FlaxAlbertForTokenClassification", "FlaxAlbertModel", "FlaxAlbertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_albert import ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, AlbertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_albert import AlbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_albert_fast import AlbertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_albert import ( ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, AlbertPreTrainedModel, load_tf_weights_in_albert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_albert import ( TF_ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFAlbertForMaskedLM, TFAlbertForMultipleChoice, TFAlbertForPreTraining, TFAlbertForQuestionAnswering, TFAlbertForSequenceClassification, TFAlbertForTokenClassification, TFAlbertMainLayer, TFAlbertModel, TFAlbertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, FlaxAlbertPreTrainedModel, ) else: import sys _UpperCamelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
645
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCamelCase : str = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys _UpperCamelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
645
1
'''simple docstring''' import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class _lowerCAmelCase : """simple docstring""" def __init__( self : Tuple , __snake_case : Dict , __snake_case : Optional[int]=13 , __snake_case : Union[str, Any]=7 , __snake_case : List[Any]=True , __snake_case : Tuple=True , __snake_case : Optional[int]=True , __snake_case : List[Any]=True , __snake_case : Tuple=99 , __snake_case : List[Any]=32 , __snake_case : int=5 , __snake_case : Tuple=4 , __snake_case : Optional[int]=4 , __snake_case : Optional[int]="gelu" , __snake_case : Optional[int]=0.0 , __snake_case : Optional[int]=0.1 , __snake_case : Any=True , __snake_case : List[Any]=5_12 , __snake_case : Tuple=16 , __snake_case : Dict=2 , __snake_case : str=0.02 , __snake_case : Dict=3 , __snake_case : Optional[Any]=4 , __snake_case : Tuple=None , )-> List[str]: snake_case = parent snake_case = batch_size snake_case = seq_length snake_case = is_training snake_case = use_input_mask snake_case = use_token_type_ids snake_case = use_labels snake_case = vocab_size snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_multiple_size snake_case = hidden_act snake_case = hidden_dropout snake_case = attention_dropout snake_case = weight_tying snake_case = max_position_embeddings snake_case = type_vocab_size snake_case = type_sequence_label_size snake_case = initializer_range snake_case = num_labels snake_case = num_choices snake_case = scope def lowerCAmelCase ( self : List[Any] )-> Any: snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case = None if self.use_input_mask: snake_case = random_attention_mask([self.batch_size, self.seq_length] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case = self.get_config() return config, input_ids, input_mask, token_labels def lowerCAmelCase ( self : List[Any] )-> List[str]: return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , ) def lowerCAmelCase ( self : List[Any] )-> List[str]: snake_case , snake_case , snake_case , snake_case = self.prepare_config_and_inputs() snake_case = True return config, input_ids, input_mask, token_labels def lowerCAmelCase ( self : Tuple , __snake_case : int , __snake_case : List[Any] , __snake_case : Tuple )-> Any: snake_case = GPTNeoXJapaneseModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) snake_case = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase ( self : Optional[Any] , __snake_case : Optional[int] , __snake_case : str , __snake_case : int )-> List[Any]: snake_case = True snake_case = GPTNeoXJapaneseModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase ( self : Optional[Any] , __snake_case : Tuple , __snake_case : str , __snake_case : Tuple , __snake_case : List[Any] )-> List[Any]: snake_case = GPTNeoXJapaneseForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase ( self : List[str] , __snake_case : Any , __snake_case : Tuple , __snake_case : Any )-> Any: snake_case = True snake_case = GPTNeoXJapaneseForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() # first forward pass snake_case = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) snake_case = output_from_no_past["""hidden_states"""][0] snake_case = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["""hidden_states"""][0] # select random slice snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1e-3 ) ) def lowerCAmelCase ( self : Tuple )-> Any: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case , snake_case = config_and_inputs snake_case = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( A__ , A__ , unittest.TestCase ): """simple docstring""" snake_case_ = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () snake_case_ = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () snake_case_ = ( {'feature-extraction': GPTNeoXJapaneseModel, 'text-generation': GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def lowerCAmelCase ( self : Dict )-> Dict: snake_case = GPTNeoXJapaneseModelTester(self ) snake_case = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=37 ) def lowerCAmelCase ( self : int )-> str: self.config_tester.run_common_tests() def lowerCAmelCase ( self : Tuple )-> Dict: snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def lowerCAmelCase ( self : Optional[int] )-> List[str]: snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def lowerCAmelCase ( self : Dict )-> Dict: snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() snake_case = None self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def lowerCAmelCase ( self : List[Any] )-> Dict: snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def lowerCAmelCase ( self : Union[str, Any] )-> List[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowerCAmelCase__ ) @slow def lowerCAmelCase ( self : List[str] )-> List[str]: snake_case = """abeja/gpt-neox-japanese-2.7b""" snake_case = ["""データサイエンティストとは、""", """100年後に必要とされる会社は、""", """フルリモートの環境で働くために必要なことは、""", """国境の長いトンネルを抜けると""", """美味しい日本食といえば、"""] snake_case = [ """データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。""", """100年後に必要とされる会社は、「人」が中心の会社です。""", """フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。""", """国境の長いトンネルを抜けると、そこは雪国だった。""", """美味しい日本食といえば、やっぱりお寿司ですよね。""", ] snake_case = GPTNeoXJapaneseTokenizer.from_pretrained(lowerCAmelCase__ ) snake_case = GPTNeoXJapaneseForCausalLM.from_pretrained(lowerCAmelCase__ ) snake_case = [] for prompt in prompts: snake_case = tokenizer(lowerCAmelCase__ , return_tensors="""pt""" ).input_ids snake_case = model.generate(lowerCAmelCase__ , max_length=50 ) snake_case = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ )
369
'''simple docstring''' import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : str ) -> Optional[int]: '''simple docstring''' debug_launcher(test_script.main ) def snake_case__ ( self : Any ) -> Optional[Any]: '''simple docstring''' debug_launcher(test_ops.main )
98
0
import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = (IPNDMScheduler,) UpperCamelCase_ = (('num_inference_steps', 5_0),) def UpperCAmelCase_ ( self : List[str] ,**lowerCAmelCase__ : Union[str, Any] ) -> int: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = {"num_train_timesteps": 10_00} config.update(**lowerCAmelCase__ ) return config def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Tuple=0 ,**lowerCAmelCase__ : int ) -> int: '''simple docstring''' lowerCAmelCase_ : str = dict(self.forward_default_kwargs ) lowerCAmelCase_ : Any = kwargs.pop("num_inference_steps" ,lowerCAmelCase__ ) lowerCAmelCase_ : Dict = self.dummy_sample lowerCAmelCase_ : int = 0.1 * sample lowerCAmelCase_ : Any = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowerCAmelCase_ : int = self.get_scheduler_config(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(lowerCAmelCase__ ) # copy over dummy past residuals lowerCAmelCase_ : Union[str, Any] = dummy_past_residuals[:] if time_step is None: lowerCAmelCase_ : Tuple = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = scheduler_class.from_pretrained(lowerCAmelCase__ ) new_scheduler.set_timesteps(lowerCAmelCase__ ) # copy over dummy past residuals lowerCAmelCase_ : Any = dummy_past_residuals[:] lowerCAmelCase_ : List[Any] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample lowerCAmelCase_ : Optional[Any] = new_scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" lowerCAmelCase_ : Any = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample lowerCAmelCase_ : Union[str, Any] = new_scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def UpperCAmelCase_ ( self : List[Any] ) -> Tuple: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any]=0 ,**lowerCAmelCase__ : List[str] ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = dict(self.forward_default_kwargs ) lowerCAmelCase_ : Dict = kwargs.pop("num_inference_steps" ,lowerCAmelCase__ ) lowerCAmelCase_ : str = self.dummy_sample lowerCAmelCase_ : List[str] = 0.1 * sample lowerCAmelCase_ : Optional[int] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowerCAmelCase_ : List[str] = self.get_scheduler_config() lowerCAmelCase_ : Union[str, Any] = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(lowerCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) lowerCAmelCase_ : List[Any] = dummy_past_residuals[:] if time_step is None: lowerCAmelCase_ : str = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__ ) lowerCAmelCase_ : int = scheduler_class.from_pretrained(lowerCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) lowerCAmelCase_ : Any = dummy_past_residuals[:] lowerCAmelCase_ : Optional[int] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample lowerCAmelCase_ : List[str] = new_scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" lowerCAmelCase_ : Union[str, Any] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample lowerCAmelCase_ : Optional[int] = new_scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def UpperCAmelCase_ ( self : int ,**lowerCAmelCase__ : int ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Any = self.scheduler_classes[0] lowerCAmelCase_ : Optional[int] = self.get_scheduler_config(**lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = scheduler_class(**lowerCAmelCase__ ) lowerCAmelCase_ : str = 10 lowerCAmelCase_ : int = self.dummy_model() lowerCAmelCase_ : Union[str, Any] = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase_ : Any = model(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase_ : int = model(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : int = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ).prev_sample return sample def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) lowerCAmelCase_ : Tuple = kwargs.pop("num_inference_steps" ,lowerCAmelCase__ ) for scheduler_class in self.scheduler_classes: lowerCAmelCase_ : Tuple = self.get_scheduler_config() lowerCAmelCase_ : List[Any] = scheduler_class(**lowerCAmelCase__ ) lowerCAmelCase_ : str = self.dummy_sample lowerCAmelCase_ : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowerCAmelCase__ ,"set_timesteps" ): scheduler.set_timesteps(lowerCAmelCase__ ) elif num_inference_steps is not None and not hasattr(lowerCAmelCase__ ,"set_timesteps" ): lowerCAmelCase_ : Optional[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowerCAmelCase_ : List[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] lowerCAmelCase_ : List[Any] = dummy_past_residuals[:] lowerCAmelCase_ : Optional[int] = scheduler.timesteps[5] lowerCAmelCase_ : List[str] = scheduler.timesteps[6] lowerCAmelCase_ : List[str] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample lowerCAmelCase_ : Union[str, Any] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) lowerCAmelCase_ : int = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample lowerCAmelCase_ : List[Any] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' for timesteps in [1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__ ,time_step=lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' for t, num_inference_steps in zip([1, 5, 10] ,[10, 50, 1_00] ): self.check_over_forward(num_inference_steps=lowerCAmelCase__ ,time_step=lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : Optional[int] = self.full_loop() lowerCAmelCase_ : Dict = torch.mean(torch.abs(lowerCAmelCase__ ) ) assert abs(result_mean.item() - 2_54_05_29 ) < 10
683
from math import sqrt def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[int] = 0 for i in range(1 , int(sqrt(snake_case__) + 1)): if n % i == 0 and i != sqrt(snake_case__): total += i + n // i elif i == sqrt(snake_case__): total += i return total - n def UpperCamelCase ( snake_case__ = 1_00_00): lowerCAmelCase_ : int = sum( i for i in range(1 , snake_case__) if sum_of_divisors(sum_of_divisors(snake_case__)) == i and sum_of_divisors(snake_case__) != i) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
683
1
from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake UpperCamelCase__ =numpy.array([0, 0]) UpperCamelCase__ =numpy.array([0.5, 0.866_0254]) UpperCamelCase__ =numpy.array([1, 0]) UpperCamelCase__ =[VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : List[Any] = initial_vectors for _ in range(__lowerCamelCase ): _SCREAMING_SNAKE_CASE : int = iteration_step(__lowerCamelCase ) return vectors def lowerCamelCase__ (__lowerCamelCase ): _SCREAMING_SNAKE_CASE : Any = [] for i, start_vector in enumerate(vectors[:-1] ): _SCREAMING_SNAKE_CASE : Any = vectors[i + 1] new_vectors.append(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3, 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = numpy.radians(__lowerCamelCase ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[str] = numpy.cos(__lowerCamelCase ), numpy.sin(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = numpy.array(((c, -s), (s, c)) ) return numpy.dot(__lowerCamelCase, __lowerCamelCase ) def lowerCamelCase__ (__lowerCamelCase ): _SCREAMING_SNAKE_CASE : Dict = plt.gca() axes.set_aspect("equal" ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : int = zip(*__lowerCamelCase ) plt.plot(__lowerCamelCase, __lowerCamelCase ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ =iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
249
"""simple docstring""" import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class a ( lowerCAmelCase_ ): def __init__( self : Union[str, Any] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : str=None , __lowerCAmelCase : List[Any]=True , __lowerCAmelCase : Union[str, Any]=None , **__lowerCAmelCase : Dict ): _UpperCAmelCase = parent _UpperCAmelCase = config_class _UpperCAmelCase = has_text_modality _UpperCAmelCase = kwargs _UpperCAmelCase = common_properties def lowerCAmelCase_ ( self : Dict ): _UpperCAmelCase = self.config_class(**self.inputs_dict ) _UpperCAmelCase = ( ["""hidden_size""", """num_attention_heads""", """num_hidden_layers"""] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["""vocab_size"""] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(__lowerCAmelCase , __lowerCAmelCase ) , msg=f'''`{prop}` does not exist''' ) # Test that config has the common properties as setter for idx, name in enumerate(__lowerCAmelCase ): try: setattr(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) self.parent.assertEqual( getattr(__lowerCAmelCase , __lowerCAmelCase ) , __lowerCAmelCase , msg=f'''`{name} value {idx} expected, but was {getattr(__lowerCAmelCase , __lowerCAmelCase )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(__lowerCAmelCase ): try: _UpperCAmelCase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(__lowerCAmelCase , __lowerCAmelCase ) , __lowerCAmelCase , msg=f'''`{name} value {idx} expected, but was {getattr(__lowerCAmelCase , __lowerCAmelCase )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def lowerCAmelCase_ ( self : Optional[Any] ): _UpperCAmelCase = self.config_class(**self.inputs_dict ) _UpperCAmelCase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , __lowerCAmelCase ) def lowerCAmelCase_ ( self : Union[str, Any] ): _UpperCAmelCase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase = os.path.join(__lowerCAmelCase , """config.json""" ) config_first.to_json_file(__lowerCAmelCase ) _UpperCAmelCase = self.config_class.from_json_file(__lowerCAmelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def lowerCAmelCase_ ( self : List[str] ): _UpperCAmelCase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(__lowerCAmelCase ) _UpperCAmelCase = self.config_class.from_pretrained(__lowerCAmelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def lowerCAmelCase_ ( self : Optional[Any] ): _UpperCAmelCase = self.config_class(**self.inputs_dict ) _UpperCAmelCase = """test""" with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) config_first.save_pretrained(__lowerCAmelCase ) _UpperCAmelCase = self.config_class.from_pretrained(__lowerCAmelCase , subfolder=__lowerCAmelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def lowerCAmelCase_ ( self : List[str] ): _UpperCAmelCase = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) _UpperCAmelCase = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def lowerCAmelCase_ ( self : List[str] ): if self.config_class.is_composition: return _UpperCAmelCase = self.config_class() self.parent.assertIsNotNone(__lowerCAmelCase ) def lowerCAmelCase_ ( self : Any ): _UpperCAmelCase = copy.deepcopy(__lowerCAmelCase ) _UpperCAmelCase = self.config_class(**__lowerCAmelCase ) _UpperCAmelCase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("""torch_dtype""", config.torch_dtype, torch.floataa) ) elif getattr(__lowerCAmelCase , __lowerCAmelCase ) != value: wrong_values.append((key, getattr(__lowerCAmelCase , __lowerCAmelCase ), value) ) if len(__lowerCAmelCase ) > 0: _UpperCAmelCase = """\n""".join([f'''- {v[0]}: got {v[1]} instead of {v[2]}''' for v in wrong_values] ) raise ValueError(f'''The following keys were not properly set in the config:\n{errors}''' ) def lowerCAmelCase_ ( self : Optional[Any] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
277
0
'''simple docstring''' import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format='''%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s''', datefmt='''%Y-%m-%d %H:%M:%S''', level=os.environ.get('''LOGLEVEL''', '''INFO''').upper(), stream=sys.stdout, ) a__ = logging.getLogger(__name__) a__ = {'facebook/bart-base': BartForConditionalGeneration} a__ = {'facebook/bart-base': BartTokenizer} def snake_case__ ( ) -> Tuple: '''simple docstring''' snake_case__ = argparse.ArgumentParser(description="""Export Bart model + Beam Search to ONNX graph.""" ) parser.add_argument( """--validation_file""" , type=lowerCamelCase_ , default=lowerCamelCase_ , help="""A csv or a json file containing the validation data.""" ) parser.add_argument( """--max_length""" , type=lowerCamelCase_ , default=5 , help="""The maximum total input sequence length after tokenization.""" , ) parser.add_argument( """--num_beams""" , type=lowerCamelCase_ , default=lowerCamelCase_ , help=( """Number of beams to use for evaluation. This argument will be """ """passed to ``model.generate``, which is used during ``evaluate`` and ``predict``.""" ) , ) parser.add_argument( """--model_name_or_path""" , type=lowerCamelCase_ , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=lowerCamelCase_ , ) parser.add_argument( """--config_name""" , type=lowerCamelCase_ , default=lowerCamelCase_ , help="""Pretrained config name or path if not the same as model_name""" , ) parser.add_argument( """--device""" , type=lowerCamelCase_ , default="""cpu""" , help="""Device where the model will be run""" , ) parser.add_argument("""--output_file_path""" , type=lowerCamelCase_ , default=lowerCamelCase_ , help="""Where to store the final ONNX file.""" ) snake_case__ = parser.parse_args() return args def snake_case__ ( a , a="cpu" ) -> str: '''simple docstring''' snake_case__ = model_dict[model_name].from_pretrained(lowerCamelCase_ ).to(lowerCamelCase_ ) snake_case__ = tokenizer_dict[model_name].from_pretrained(lowerCamelCase_ ) if model_name in ["facebook/bart-base"]: snake_case__ = 0 snake_case__ = None snake_case__ = 0 return huggingface_model, tokenizer def snake_case__ ( a , a , a , a , a ) -> int: '''simple docstring''' model.eval() snake_case__ = None snake_case__ = torch.jit.script(BARTBeamSearchGenerator(lowerCamelCase_ ) ) with torch.no_grad(): snake_case__ = '''My friends are cool but they eat too many carbs.''' snake_case__ = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=1024 , return_tensors="""pt""" ).to(model.device ) snake_case__ = model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , num_beams=lowerCamelCase_ , max_length=lowerCamelCase_ , early_stopping=lowerCamelCase_ , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( lowerCamelCase_ , ( inputs["""input_ids"""], inputs["""attention_mask"""], num_beams, max_length, model.config.decoder_start_token_id, ) , lowerCamelCase_ , opset_version=14 , input_names=["""input_ids""", """attention_mask""", """num_beams""", """max_length""", """decoder_start_token_id"""] , output_names=["""output_ids"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """seq"""}, """output_ids""": {0: """batch""", 1: """seq_out"""}, } , example_outputs=lowerCamelCase_ , ) logger.info("""Model exported to {}""".format(lowerCamelCase_ ) ) snake_case__ = remove_dup_initializers(os.path.abspath(lowerCamelCase_ ) ) logger.info("""Deduplicated and optimized model written to {}""".format(lowerCamelCase_ ) ) snake_case__ = onnxruntime.InferenceSession(lowerCamelCase_ ) snake_case__ = ort_sess.run( lowerCamelCase_ , { """input_ids""": inputs["""input_ids"""].cpu().numpy(), """attention_mask""": inputs["""attention_mask"""].cpu().numpy(), """num_beams""": np.array(lowerCamelCase_ ), """max_length""": np.array(lowerCamelCase_ ), """decoder_start_token_id""": np.array(model.config.decoder_start_token_id ), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1e-3 , atol=1e-3 ) logger.info("""Model outputs from torch and ONNX Runtime are similar.""" ) logger.info("""Success.""" ) def snake_case__ ( ) -> Dict: '''simple docstring''' snake_case__ = parse_args() snake_case__ = 5 snake_case__ = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.setLevel(logging.INFO ) transformers.utils.logging.set_verbosity_error() snake_case__ = torch.device(args.device ) snake_case__ = load_model_tokenizer(args.model_name_or_path , lowerCamelCase_ ) if model.config.decoder_start_token_id is None: raise ValueError("""Make sure that `config.decoder_start_token_id` is correctly defined""" ) model.to(lowerCamelCase_ ) if args.max_length: snake_case__ = args.max_length if args.num_beams: snake_case__ = args.num_beams if args.output_file_path: snake_case__ = args.output_file_path else: snake_case__ = '''BART.onnx''' logger.info("""Exporting model to ONNX""" ) export_and_validate_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) if __name__ == "__main__": main()
702
'''simple docstring''' from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging a__ = logging.get_logger(__name__) a__ = { '''huggingface/informer-tourism-monthly''': ( '''https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json''' ), # See all Informer models at https://huggingface.co/models?filter=informer } class __magic_name__( __lowerCAmelCase ): UpperCAmelCase_ : str = """informer""" UpperCAmelCase_ : Any = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : Optional[Any] , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : str = "student_t" , __UpperCamelCase : str = "nll" , __UpperCamelCase : int = 1 , __UpperCamelCase : List[int] = None , __UpperCamelCase : Optional[Union[str, bool]] = "mean" , __UpperCamelCase : int = 0 , __UpperCamelCase : int = 0 , __UpperCamelCase : int = 0 , __UpperCamelCase : int = 0 , __UpperCamelCase : Optional[List[int]] = None , __UpperCamelCase : Optional[List[int]] = None , __UpperCamelCase : int = 6_4 , __UpperCamelCase : int = 3_2 , __UpperCamelCase : int = 3_2 , __UpperCamelCase : int = 2 , __UpperCamelCase : int = 2 , __UpperCamelCase : int = 2 , __UpperCamelCase : int = 2 , __UpperCamelCase : bool = True , __UpperCamelCase : str = "gelu" , __UpperCamelCase : float = 0.05 , __UpperCamelCase : float = 0.1 , __UpperCamelCase : float = 0.1 , __UpperCamelCase : float = 0.1 , __UpperCamelCase : float = 0.1 , __UpperCamelCase : int = 1_0_0 , __UpperCamelCase : float = 0.02 , __UpperCamelCase : Union[str, Any]=True , __UpperCamelCase : str = "prob" , __UpperCamelCase : int = 5 , __UpperCamelCase : bool = True , **__UpperCamelCase : Any , ): '''simple docstring''' snake_case__ = prediction_length snake_case__ = context_length or prediction_length snake_case__ = distribution_output snake_case__ = loss snake_case__ = input_size snake_case__ = num_time_features snake_case__ = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] snake_case__ = scaling snake_case__ = num_dynamic_real_features snake_case__ = num_static_real_features snake_case__ = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(__UpperCamelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) snake_case__ = cardinality else: snake_case__ = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(__UpperCamelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) snake_case__ = embedding_dimension else: snake_case__ = [min(5_0 , (cat + 1) // 2 ) for cat in self.cardinality] snake_case__ = num_parallel_samples # Transformer architecture configuration snake_case__ = input_size * len(self.lags_sequence ) + self._number_of_features snake_case__ = d_model snake_case__ = encoder_attention_heads snake_case__ = decoder_attention_heads snake_case__ = encoder_ffn_dim snake_case__ = decoder_ffn_dim snake_case__ = encoder_layers snake_case__ = decoder_layers snake_case__ = dropout snake_case__ = attention_dropout snake_case__ = activation_dropout snake_case__ = encoder_layerdrop snake_case__ = decoder_layerdrop snake_case__ = activation_function snake_case__ = init_std snake_case__ = use_cache # Informer snake_case__ = attention_type snake_case__ = sampling_factor snake_case__ = distil super().__init__(is_encoder_decoder=__UpperCamelCase , **__UpperCamelCase ) @property def __lowerCAmelCase( self : str ): '''simple docstring''' return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
566
0
"""simple docstring""" def lowercase (_lowerCAmelCase , _lowerCAmelCase ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) __lowerCAmelCase = (boundary[1] - boundary[0]) / steps __lowerCAmelCase = boundary[0] __lowerCAmelCase = boundary[1] __lowerCAmelCase = make_points(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowerCAmelCase = 0.0 y += (h / 2.0) * f(_lowerCAmelCase ) for i in x_i: # print(i) y += h * f(_lowerCAmelCase ) y += (h / 2.0) * f(_lowerCAmelCase ) return y def lowercase (_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): __lowerCAmelCase = a + h while x < (b - h): yield x __lowerCAmelCase = x + h def lowercase (_lowerCAmelCase ): # enter your function here __lowerCAmelCase = (x - 0) * (x - 0) return y def lowercase (): __lowerCAmelCase = 0.0 # Lower bound of integration __lowerCAmelCase = 1.0 # Upper bound of integration __lowerCAmelCase = 10.0 # define number of steps or resolution __lowerCAmelCase = [a, b] # define boundary of integration __lowerCAmelCase = method_a(_lowerCAmelCase , _lowerCAmelCase ) print(f"""y = {y}""" ) if __name__ == "__main__": main()
465
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available SCREAMING_SNAKE_CASE_ = { '''configuration_rag''': ['''RagConfig'''], '''retrieval_rag''': ['''RagRetriever'''], '''tokenization_rag''': ['''RagTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ '''RagModel''', '''RagPreTrainedModel''', '''RagSequenceForGeneration''', '''RagTokenForGeneration''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ '''TFRagModel''', '''TFRagPreTrainedModel''', '''TFRagSequenceForGeneration''', '''TFRagTokenForGeneration''', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
465
1
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class UpperCAmelCase ( unittest.TestCase): """simple docstring""" def __init__( self : int , UpperCamelCase__ : str , UpperCamelCase__ : Any=13 , UpperCamelCase__ : Optional[int]=7 , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : int=True , UpperCamelCase__ : Dict=True , UpperCamelCase__ : int=True , UpperCamelCase__ : Union[str, Any]=99 , UpperCamelCase__ : str=32 , UpperCamelCase__ : str=5 , UpperCamelCase__ : List[Any]=4 , UpperCamelCase__ : List[Any]=37 , UpperCamelCase__ : int="gelu" , UpperCamelCase__ : Optional[int]=0.1 , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : Dict=512 , UpperCamelCase__ : Any=16 , UpperCamelCase__ : int=2 , UpperCamelCase__ : Optional[Any]=0.02 , UpperCamelCase__ : str=4 , ) -> Tuple: _UpperCamelCase =parent _UpperCamelCase =batch_size _UpperCamelCase =seq_length _UpperCamelCase =is_training _UpperCamelCase =use_attention_mask _UpperCamelCase =use_token_type_ids _UpperCamelCase =use_labels _UpperCamelCase =vocab_size _UpperCamelCase =hidden_size _UpperCamelCase =num_hidden_layers _UpperCamelCase =num_attention_heads _UpperCamelCase =intermediate_size _UpperCamelCase =hidden_act _UpperCamelCase =hidden_dropout_prob _UpperCamelCase =attention_probs_dropout_prob _UpperCamelCase =max_position_embeddings _UpperCamelCase =type_vocab_size _UpperCamelCase =type_sequence_label_size _UpperCamelCase =initializer_range _UpperCamelCase =num_choices def UpperCamelCase__ ( self : int ) -> Optional[int]: _UpperCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase =None if self.use_attention_mask: _UpperCamelCase =random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase =None if self.use_token_type_ids: _UpperCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase =AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCamelCase__ ( self : int ) -> List[str]: _UpperCamelCase =self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase =config_and_inputs _UpperCamelCase ={'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class UpperCAmelCase ( lowercase_ , unittest.TestCase): """simple docstring""" lowerCAmelCase_ = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase__ ( self : Tuple ) -> Optional[int]: _UpperCamelCase =FlaxAlbertModelTester(self ) @slow def UpperCamelCase__ ( self : List[str] ) -> Union[str, Any]: for model_class_name in self.all_model_classes: _UpperCamelCase =model_class_name.from_pretrained('''albert-base-v2''' ) _UpperCamelCase =model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase__ ) @require_flax class UpperCAmelCase ( unittest.TestCase): """simple docstring""" @slow def UpperCamelCase__ ( self : Optional[int] ) -> str: _UpperCamelCase =FlaxAlbertModel.from_pretrained('''albert-base-v2''' ) _UpperCamelCase =np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) _UpperCamelCase =np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCamelCase =model(UpperCamelCase__ , attention_mask=UpperCamelCase__ )[0] _UpperCamelCase =(1, 11, 768) self.assertEqual(output.shape , UpperCamelCase__ ) _UpperCamelCase =np.array( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCamelCase__ , atol=1E-4 ) )
702
'''simple docstring''' from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar __lowerCamelCase : Optional[int] = TypeVar('T') class UpperCAmelCase ( Generic[T]): """simple docstring""" lowerCAmelCase_ = 42 # Cache store of keys lowerCAmelCase_ = 42 # References of the keys in cache lowerCAmelCase_ = 10 # Maximum capacity of cache def __init__( self : Union[str, Any] , UpperCamelCase__ : int ) -> None: _UpperCamelCase =deque() _UpperCamelCase =set() if not n: _UpperCamelCase =sys.maxsize elif n < 0: raise ValueError('''n should be an integer greater than 0.''' ) else: _UpperCamelCase =n def UpperCamelCase__ ( self : List[str] , UpperCamelCase__ : T ) -> None: if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: _UpperCamelCase =self.dq_store.pop() self.key_reference.remove(UpperCamelCase__ ) else: self.dq_store.remove(UpperCamelCase__ ) self.dq_store.appendleft(UpperCamelCase__ ) self.key_reference.add(UpperCamelCase__ ) def UpperCamelCase__ ( self : Union[str, Any] ) -> None: for k in self.dq_store: print(UpperCamelCase__ ) def __repr__( self : str ) -> str: return F'''LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}''' if __name__ == "__main__": import doctest doctest.testmod() __lowerCamelCase : LRUCache[str | int] = LRUCache(4) lru_cache.refer('A') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('A') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
271
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) a : Optional[Any] = {'''configuration_plbart''': ['''PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PLBartConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[int] = ['''PLBartTokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ '''PLBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PLBartForCausalLM''', '''PLBartForConditionalGeneration''', '''PLBartForSequenceClassification''', '''PLBartModel''', '''PLBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys a : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
555
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def _UpperCamelCase ( _A , _A=False ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = OmegaConf.load(_A ) if display: print(yaml.dump(OmegaConf.to_container(_A ) ) ) return config def _UpperCamelCase ( _A , _A=None , _A=None ) -> List[Any]: """simple docstring""" if conf_path is None: _UpperCAmelCase = """./model_checkpoints/vqgan_only.yaml""" _UpperCAmelCase = load_config(_A , display=_A ) _UpperCAmelCase = VQModel(**config.model.params ) if ckpt_path is None: _UpperCAmelCase = """./model_checkpoints/vqgan_only.pt""" _UpperCAmelCase = torch.load(_A , map_location=_A ) if ".ckpt" in ckpt_path: _UpperCAmelCase = sd["""state_dict"""] model.load_state_dict(_A , strict=_A ) model.to(_A ) del sd return model def _UpperCamelCase ( _A , _A ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase = model.encode(_A ) print(F"""VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}""" ) _UpperCAmelCase = model.decode(_A ) return xrec def _UpperCamelCase ( _A , _A=False ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase ,_UpperCAmelCase = string.rsplit(""".""" , 1 ) if reload: _UpperCAmelCase = importlib.import_module(_A ) importlib.reload(_A ) return getattr(importlib.import_module(_A , package=_A ) , cls ) def _UpperCamelCase ( _A ) -> str: """simple docstring""" if "target" not in config: raise KeyError("""Expected key `target` to instantiate.""" ) return get_obj_from_str(config["""target"""] )(**config.get("""params""" , {} ) ) def _UpperCamelCase ( _A , _A , _A=True , _A=True ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = instantiate_from_config(_A ) if sd is not None: model.load_state_dict(_A ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def _UpperCamelCase ( _A , _A , _A , _A ) -> Optional[Any]: """simple docstring""" if ckpt: _UpperCAmelCase = torch.load(_A , map_location="""cpu""" ) _UpperCAmelCase = pl_sd["""global_step"""] print(F"""loaded model from global step {global_step}.""" ) else: _UpperCAmelCase = {"""state_dict""": None} _UpperCAmelCase = None _UpperCAmelCase = load_model_from_config(config.model , pl_sd["""state_dict"""] , gpu=_A , eval_mode=_A )["""model"""] return model, global_step
555
1
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def _lowerCamelCase ( A_ : Optional[Any] ) -> str: '''simple docstring''' if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: UpperCamelCase__ : Any =k.replace(lowerCAmelCase__ , lowerCAmelCase__ ) if k.startswith("encoder" ): UpperCamelCase__ : List[Any] =k.replace(".attn" , ".self_attn" ) UpperCamelCase__ : Dict =k.replace("norm1" , "self_attn_layer_norm" ) UpperCamelCase__ : Union[str, Any] =k.replace("norm2" , "final_layer_norm" ) elif k.startswith("decoder" ): UpperCamelCase__ : List[Any] =k.replace("norm1" , "self_attn_layer_norm" ) UpperCamelCase__ : Any =k.replace("norm2" , "encoder_attn_layer_norm" ) UpperCamelCase__ : Tuple =k.replace("norm3" , "final_layer_norm" ) return k def _lowerCamelCase ( A_ : Dict ) -> Tuple: '''simple docstring''' UpperCamelCase__ : Any =[ "model.encoder.layernorm_embedding.weight", "model.encoder.layernorm_embedding.bias", "model.decoder.layernorm_embedding.weight", "model.decoder.layernorm_embedding.bias", ] for k in keys: UpperCamelCase__ : Union[str, Any] =sd.pop(lowerCAmelCase__ ) UpperCamelCase__ : Tuple =k.replace("layernorm_embedding" , "layer_norm" ) assert new_k not in sd UpperCamelCase__ : Tuple =v __UpperCAmelCase = ["""START"""] @torch.no_grad() def _lowerCamelCase ( A_ : Tuple , A_ : str , A_ : Optional[int] ) -> Any: '''simple docstring''' UpperCamelCase__ : List[str] =torch.load(lowerCAmelCase__ , map_location="cpu" ) UpperCamelCase__ : int =model["model"] UpperCamelCase__ : str =BlenderbotConfig.from_json_file(lowerCAmelCase__ ) UpperCamelCase__ : str =BlenderbotForConditionalGeneration(lowerCAmelCase__ ) UpperCamelCase__ : Tuple =m.model.state_dict().keys() UpperCamelCase__ : List[Any] =[] UpperCamelCase__ : int ={} for k, v in sd.items(): if k in IGNORE_KEYS: continue UpperCamelCase__ : Union[str, Any] =rename_state_dict_key(lowerCAmelCase__ ) if new_k not in valid_keys: failures.append([k, new_k] ) else: UpperCamelCase__ : List[Any] =v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(lowerCAmelCase__ ) m.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) m.half() m.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) __UpperCAmelCase = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
714
def _lowerCamelCase ( A_ : int , A_ : int ) -> int: '''simple docstring''' return x if y == 0 else greatest_common_divisor(A_ , x % y ) def _lowerCamelCase ( A_ : int , A_ : int ) -> int: '''simple docstring''' return (x * y) // greatest_common_divisor(A_ , A_ ) def _lowerCamelCase ( A_ : int = 2_0 ) -> int: '''simple docstring''' UpperCamelCase__ : Any =1 for i in range(1 , n + 1 ): UpperCamelCase__ : int =lcm(A_ , A_ ) return g if __name__ == "__main__": print(F"""{solution() = }""")
582
0
"""simple docstring""" lowerCamelCase_ = 9.80665 def snake_case ( A__ ,A__ ,A__ = g ): if fluid_density <= 0: raise ValueError("Impossible fluid density" ) if volume < 0: raise ValueError("Impossible Object volume" ) if gravity <= 0: raise ValueError("Impossible Gravity" ) return fluid_density * gravity * volume if __name__ == "__main__": import doctest # run doctest doctest.testmod()
95
from collections import deque def A__ ( __A : Optional[Any] ) ->Tuple: __A =len(__A ) __A =deque() __A =[False for _ in range(__A )] __A =[-1 for _ in range(__A )] __A =index_of[:] def strong_connect(__A : Union[str, Any] , __A : int , __A : Optional[int] ): __A =index # the number when this node is seen __A =index # lowest rank node reachable from here index += 1 stack.append(__A ) __A =True for w in g[v]: if index_of[w] == -1: __A =strong_connect(__A , __A , __A ) __A =( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: __A =( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: __A =[] __A =stack.pop() __A =False component.append(__A ) while w != v: __A =stack.pop() __A =False component.append(__A ) components.append(__A ) return index __A =[] for v in range(__A ): if index_of[v] == -1: strong_connect(__A , 0 , __A ) return components def A__ ( __A : List[Any] , __A : Optional[int] ) ->Tuple: __A =[[] for _ in range(__A )] for u, v in edges: g[u].append(__A ) return g if __name__ == "__main__": # Test _lowerCamelCase : int = 7 _lowerCamelCase : Tuple = [0, 0, 1, 2, 3, 3, 4, 4, 6] _lowerCamelCase : Dict = [1, 3, 2, 0, 1, 4, 5, 6, 5] _lowerCamelCase : Optional[Any] = [(u, v) for u, v in zip(source, target)] _lowerCamelCase : Dict = create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
184
0
"""simple docstring""" import math def a__ ( snake_case__ ) -> bool: assert isinstance(snake_case__ , snake_case__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False lowerCamelCase = range(3 , int(math.sqrt(snake_case__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def a__ ( snake_case__ , snake_case__=1 , **snake_case__ ) -> Optional[Any]: lowerCamelCase = factor * value lowerCamelCase = value while not is_prime(snake_case__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **snake_case__ ) return value
533
"""simple docstring""" from __future__ import annotations def a__ ( snake_case__ , snake_case__ = None , snake_case__ = None ) -> None: if start is None: lowerCamelCase = 0 if end is None: lowerCamelCase = len(snake_case__ ) - 1 if start >= end: return lowerCamelCase = (start + end) // 2 slowsort(snake_case__ , snake_case__ , snake_case__ ) slowsort(snake_case__ , mid + 1 , snake_case__ ) if sequence[end] < sequence[mid]: lowerCamelCase , lowerCamelCase = sequence[mid], sequence[end] slowsort(snake_case__ , snake_case__ , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
533
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = torch.device('cpu') def __SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowercase_ : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase_ : str = Image.open(requests.get(_UpperCamelCase , stream=_UpperCamelCase ).raw ) return im def __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): """simple docstring""" if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1703e00, 2.1107e00, -2.0811e00, 8.8685e-01, 2.4360e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9636e-01, 2.3478e-01, -1.6963e00, -1.7381e00, -8.6337e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2768e-01, -4.7429e-01, -1.0897e00, -1.0248e00, 3.5523e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5330e-01, 2.4211e-01, -6.0185e-01, -8.2789e-01, -6.0446e-02] ) def __SCREAMING_SNAKE_CASE ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowercase_ : Optional[int] = dct.pop(_UpperCamelCase ) lowercase_ : Dict = val def __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): """simple docstring""" lowercase_ : Union[str, Any] = [] for k in state_dict.keys(): lowercase_ : Dict = k if ".pwconv" in k: lowercase_ : Any = k_new.replace(".pwconv" , ".point_wise_conv" ) if ".dwconv" in k: lowercase_ : Any = k_new.replace(".dwconv" , ".depth_wise_conv" ) if ".Proj." in k: lowercase_ : Dict = k_new.replace(".Proj." , ".proj." ) if "patch_embed" in k_new: lowercase_ : Tuple = k_new.replace("patch_embed" , "swiftformer.patch_embed.patch_embedding" ) if "network" in k_new: lowercase_ : Any = k_new.split("." ) if ls[2].isdigit(): lowercase_ : List[str] = "swiftformer.encoder.network." + ls[1] + ".blocks." + ls[2] + "." + ".".join(ls[3:] ) else: lowercase_ : Optional[Any] = k_new.replace("network" , "swiftformer.encoder.network" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def __SCREAMING_SNAKE_CASE ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowercase_ : str = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size lowercase_ : Any = 1000 lowercase_ : List[Any] = "huggingface/label-files" lowercase_ : Any = "imagenet-1k-id2label.json" lowercase_ : List[str] = json.load(open(hf_hub_download(_UpperCamelCase , _UpperCamelCase , repo_type="dataset" ) , "r" ) ) lowercase_ : List[str] = {int(_UpperCamelCase ): v for k, v in idalabel.items()} lowercase_ : int = idalabel lowercase_ : Union[str, Any] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": lowercase_ : Tuple = [3, 3, 6, 4] lowercase_ : int = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": lowercase_ : List[str] = [3, 3, 9, 6] lowercase_ : List[str] = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": lowercase_ : Dict = [4, 3, 10, 5] lowercase_ : Optional[int] = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": lowercase_ : List[Any] = [4, 4, 12, 6] lowercase_ : Union[str, Any] = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("https" ): lowercase_ : Optional[Any] = torch.hub.load_state_dict_from_url(_UpperCamelCase , map_location="cpu" , check_hash=_UpperCamelCase ) else: lowercase_ : Any = torch.load(_UpperCamelCase , map_location="cpu" ) lowercase_ : Dict = checkpoint lowercase_ : Optional[int] = create_rename_keys(_UpperCamelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # load HuggingFace model lowercase_ : int = SwiftFormerForImageClassification(_UpperCamelCase ).eval() hf_model.load_state_dict(_UpperCamelCase ) # prepare test inputs lowercase_ : List[str] = prepare_img() lowercase_ : Union[str, Any] = ViTImageProcessor.from_pretrained("preprocessor_config" ) lowercase_ : List[str] = processor(images=_UpperCamelCase , return_tensors="pt" ) # compare outputs from both models lowercase_ : Any = get_expected_output(_UpperCamelCase ) lowercase_ : List[Any] = hf_model(inputs["pixel_values"] ).logits assert hf_logits.shape == torch.Size([1, 1000] ) assert torch.allclose(hf_logits[0, 0:5] , _UpperCamelCase , atol=1e-3 ) Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') UpperCamelCase__ = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
620
'''simple docstring''' import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP UpperCamelCase__ = False try: UpperCamelCase__ = _is_package_available('google.colab') except ModuleNotFoundError: pass @input.register class _UpperCAmelCase : def __init__( self : List[Any] , a : str = None , a : list = [] ): '''simple docstring''' lowercase_ : Optional[Any] = 0 lowercase_ : Dict = choices lowercase_ : List[str] = prompt if sys.platform == "win32": lowercase_ : List[Any] = "*" else: lowercase_ : Union[str, Any] = "➔ " def lowerCAmelCase__ ( self : int , a : Optional[int] , a : str = "" ): '''simple docstring''' if sys.platform != "win32": writeColor(self.choices[index] , 3_2 , a ) else: forceWrite(self.choices[index] , a ) def lowerCAmelCase__ ( self : List[Any] , a : int ): '''simple docstring''' if index == self.position: forceWrite(f""" {self.arrow_char} """ ) self.write_choice(a ) else: forceWrite(f""" {self.choices[index]}""" ) reset_cursor() def lowerCAmelCase__ ( self : str , a : Direction , a : int = 1 ): '''simple docstring''' lowercase_ : Optional[int] = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(a ) move_cursor(a , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP["up"] ) def lowerCAmelCase__ ( self : Dict ): '''simple docstring''' self.move_direction(Direction.UP ) @input.mark(KEYMAP["down"] ) def lowerCAmelCase__ ( self : List[Any] ): '''simple docstring''' self.move_direction(Direction.DOWN ) @input.mark(KEYMAP["newline"] ) def lowerCAmelCase__ ( self : Any ): '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) return self.position @input.mark(KEYMAP["interrupt"] ) def lowerCAmelCase__ ( self : str ): '''simple docstring''' move_cursor(len(self.choices ) - self.position , "DOWN" ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(a )] for number in range(1_0 )] ) def lowerCAmelCase__ ( self : int ): '''simple docstring''' lowercase_ : str = int(chr(self.current_selection ) ) lowercase_ : Tuple = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , a ) else: return else: return def lowerCAmelCase__ ( self : Tuple , a : int = 0 ): '''simple docstring''' if self.prompt: linebreak() forceWrite(self.prompt , "\n" ) if in_colab: forceWrite("Please input a choice index (starting from 0), and press enter" , "\n" ) else: forceWrite("Please select a choice using the arrow or number keys, and selecting with enter" , "\n" ) lowercase_ : Dict = default_choice for i in range(len(self.choices ) ): self.print_choice(a ) forceWrite("\n" ) move_cursor(len(self.choices ) - self.position , "UP" ) with cursor.hide(): while True: if in_colab: try: lowercase_ : List[str] = int(builtins.input() ) except ValueError: lowercase_ : List[Any] = default_choice else: lowercase_ : Dict = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , "UP" ) clear_line() self.write_choice(a , "\n" ) return choice
620
1
from ...processing_utils import ProcessorMixin class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case = ["image_processor", "feature_extractor"] snake_case = "TvltImageProcessor" snake_case = "TvltFeatureExtractor" def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ ): super().__init__(image_processor=A_ , feature_extractor=A_ ) snake_case_ = image_processor snake_case_ = feature_extractor def __call__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=False , UpperCAmelCase_=False , *UpperCAmelCase_ , **UpperCAmelCase_ , ): if images is None and audio is None: raise ValueError("You need to specify either an `images` or `audio` input to process." ) snake_case_ = None if images is not None: snake_case_ = self.image_processor(A_ , mask_pixel=A_ , *A_ , **A_ ) if images_mixed is not None: snake_case_ = self.image_processor(A_ , is_mixed=A_ , *A_ , **A_ ) if audio is not None: snake_case_ = self.feature_extractor( A_ , *A_ , sampling_rate=A_ , mask_audio=A_ , **A_ ) snake_case_ = {} if audio is not None: output_dict.update(A_ ) if images is not None: output_dict.update(A_ ) if images_mixed_dict is not None: output_dict.update(A_ ) return output_dict @property def _lowercase ( self ): snake_case_ = self.image_processor.model_input_names snake_case_ = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
703
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = """philschmid/bart-large-cnn-samsum""" snake_case = ( """This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, """ """and returns a summary of the text.""" ) snake_case = """summarizer""" snake_case = AutoTokenizer snake_case = AutoModelForSeqaSeqLM snake_case = ["""text"""] snake_case = ["""text"""] def _lowercase ( self , UpperCAmelCase_ ): return self.pre_processor(UpperCAmelCase_ , return_tensors="pt" , truncation=UpperCAmelCase_ ) def _lowercase ( self , UpperCAmelCase_ ): return self.model.generate(**UpperCAmelCase_ )[0] def _lowercase ( self , UpperCAmelCase_ ): return self.pre_processor.decode(UpperCAmelCase_ , skip_special_tokens=UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ )
420
0
'''simple docstring''' # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _SCREAMING_SNAKE_CASE (*A ) -> Optional[int]: """simple docstring""" with open(A , '''r''' ) as fh: fcntl.flock(A , fcntl.LOCK_EX ) try: print(*A ) finally: fcntl.flock(A , fcntl.LOCK_UN ) lowerCamelCase : Dict = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) lowerCamelCase : List[Any] = torch.device('cuda', local_rank) lowerCamelCase : Optional[Any] = socket.gethostname() lowerCamelCase : int = f"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank lowerCamelCase : Union[str, Any] = dist.get_rank() lowerCamelCase : str = dist.get_world_size() printflock(f"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(f"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(f"""{gpu} is broken""") raise
460
'''simple docstring''' def _SCREAMING_SNAKE_CASE (A ) -> bool: """simple docstring""" if not isinstance(A , A ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(A ) == 0: raise ValueError('''Input list must be a non empty list''' ) if len(A ) == 1: return True lowercase__ = series[1] - series[0] for index in range(len(A ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _SCREAMING_SNAKE_CASE (A ) -> float: """simple docstring""" if not isinstance(A , A ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(A ) == 0: raise ValueError('''Input list must be a non empty list''' ) lowercase__ = 0 for val in series: answer += val return answer / len(A ) if __name__ == "__main__": import doctest doctest.testmod()
460
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A__ ( A ): """simple docstring""" _lowercase : int = (PNDMScheduler,) _lowercase : List[str] = (('''num_inference_steps''', 50),) def __magic_name__ ( self : str , **A_ : Dict ): '''simple docstring''' _lowerCAmelCase : Tuple = { "num_train_timesteps": 1_0_0_0, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**A_ ) return config def __magic_name__ ( self : Dict , A_ : Union[str, Any]=0 , **A_ : List[Any] ): '''simple docstring''' _lowerCAmelCase : Tuple = dict(self.forward_default_kwargs ) _lowerCAmelCase : Union[str, Any] = kwargs.pop("num_inference_steps" , A_ ) _lowerCAmelCase : Union[str, Any] = self.dummy_sample _lowerCAmelCase : Optional[Any] = 0.1 * sample _lowerCAmelCase : Union[str, Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _lowerCAmelCase : Dict = self.get_scheduler_config(**A_ ) _lowerCAmelCase : List[Any] = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # copy over dummy past residuals _lowerCAmelCase : List[str] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A_ ) _lowerCAmelCase : Optional[Any] = scheduler_class.from_pretrained(A_ ) new_scheduler.set_timesteps(A_ ) # copy over dummy past residuals _lowerCAmelCase : List[Any] = dummy_past_residuals[:] _lowerCAmelCase : List[str] = scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample _lowerCAmelCase : Union[str, Any] = new_scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _lowerCAmelCase : Dict = scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample _lowerCAmelCase : int = new_scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def __magic_name__ ( self : str ): '''simple docstring''' pass def __magic_name__ ( self : str , A_ : Any=0 , **A_ : Optional[int] ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = dict(self.forward_default_kwargs ) _lowerCAmelCase : Tuple = kwargs.pop("num_inference_steps" , A_ ) _lowerCAmelCase : int = self.dummy_sample _lowerCAmelCase : int = 0.1 * sample _lowerCAmelCase : List[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _lowerCAmelCase : Dict = self.get_scheduler_config() _lowerCAmelCase : Optional[int] = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # copy over dummy past residuals (must be after setting timesteps) _lowerCAmelCase : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A_ ) _lowerCAmelCase : List[str] = scheduler_class.from_pretrained(A_ ) # copy over dummy past residuals new_scheduler.set_timesteps(A_ ) # copy over dummy past residual (must be after setting timesteps) _lowerCAmelCase : List[Any] = dummy_past_residuals[:] _lowerCAmelCase : Optional[Any] = scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample _lowerCAmelCase : List[str] = new_scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _lowerCAmelCase : Any = scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample _lowerCAmelCase : List[str] = new_scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def __magic_name__ ( self : Optional[int] , **A_ : Optional[Any] ): '''simple docstring''' _lowerCAmelCase : Any = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config(**A_ ) _lowerCAmelCase : Any = scheduler_class(**A_ ) _lowerCAmelCase : List[Any] = 1_0 _lowerCAmelCase : List[str] = self.dummy_model() _lowerCAmelCase : int = self.dummy_sample_deter scheduler.set_timesteps(A_ ) for i, t in enumerate(scheduler.prk_timesteps ): _lowerCAmelCase : Union[str, Any] = model(A_ , A_ ) _lowerCAmelCase : Optional[Any] = scheduler.step_prk(A_ , A_ , A_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): _lowerCAmelCase : Any = model(A_ , A_ ) _lowerCAmelCase : Optional[Any] = scheduler.step_plms(A_ , A_ , A_ ).prev_sample return sample def __magic_name__ ( self : Tuple ): '''simple docstring''' _lowerCAmelCase : Dict = dict(self.forward_default_kwargs ) _lowerCAmelCase : Optional[int] = kwargs.pop("num_inference_steps" , A_ ) for scheduler_class in self.scheduler_classes: _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Tuple = scheduler_class(**A_ ) _lowerCAmelCase : List[Any] = self.dummy_sample _lowerCAmelCase : Optional[int] = 0.1 * sample if num_inference_steps is not None and hasattr(A_ , "set_timesteps" ): scheduler.set_timesteps(A_ ) elif num_inference_steps is not None and not hasattr(A_ , "set_timesteps" ): _lowerCAmelCase : List[str] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _lowerCAmelCase : Optional[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] _lowerCAmelCase : Union[str, Any] = dummy_past_residuals[:] _lowerCAmelCase : Any = scheduler.step_prk(A_ , 0 , A_ , **A_ ).prev_sample _lowerCAmelCase : List[str] = scheduler.step_prk(A_ , 1 , A_ , **A_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _lowerCAmelCase : Dict = scheduler.step_plms(A_ , 0 , A_ , **A_ ).prev_sample _lowerCAmelCase : int = scheduler.step_plms(A_ , 1 , A_ , **A_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __magic_name__ ( self : str ): '''simple docstring''' for timesteps in [1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A_ ) def __magic_name__ ( self : int ): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=A_ ) _lowerCAmelCase : Tuple = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config(steps_offset=1 ) _lowerCAmelCase : List[Any] = scheduler_class(**A_ ) scheduler.set_timesteps(1_0 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [9_0_1, 8_5_1, 8_5_1, 8_0_1, 8_0_1, 7_5_1, 7_5_1, 7_0_1, 7_0_1, 6_5_1, 6_5_1, 6_0_1, 6_0_1, 5_0_1, 4_0_1, 3_0_1, 2_0_1, 1_0_1, 1] ) , ) def __magic_name__ ( self : Any ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001] , [0.002, 0.02] ): self.check_over_configs(beta_start=A_ , beta_end=A_ ) def __magic_name__ ( self : List[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=A_ ) def __magic_name__ ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A_ ) def __magic_name__ ( self : Optional[int] ): '''simple docstring''' for t in [1, 5, 1_0]: self.check_over_forward(time_step=A_ ) def __magic_name__ ( self : str ): '''simple docstring''' for t, num_inference_steps in zip([1, 5, 1_0] , [1_0, 5_0, 1_0_0] ): self.check_over_forward(num_inference_steps=A_ ) def __magic_name__ ( self : Dict ): '''simple docstring''' _lowerCAmelCase : Dict = 2_7 for scheduler_class in self.scheduler_classes: _lowerCAmelCase : Dict = self.dummy_sample _lowerCAmelCase : int = 0.1 * sample _lowerCAmelCase : Any = self.get_scheduler_config() _lowerCAmelCase : List[str] = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): _lowerCAmelCase : List[Any] = scheduler.step_prk(A_ , A_ , A_ ).prev_sample def __magic_name__ ( self : str ): '''simple docstring''' with self.assertRaises(A_ ): _lowerCAmelCase : Tuple = self.scheduler_classes[0] _lowerCAmelCase : Union[str, Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**A_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def __magic_name__ ( self : Optional[int] ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.full_loop() _lowerCAmelCase : Dict = torch.sum(torch.abs(A_ ) ) _lowerCAmelCase : int = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 198.1318 ) < 1E-2 assert abs(result_mean.item() - 0.2580 ) < 1E-3 def __magic_name__ ( self : Optional[Any] ): '''simple docstring''' _lowerCAmelCase : Tuple = self.full_loop(prediction_type="v_prediction" ) _lowerCAmelCase : int = torch.sum(torch.abs(A_ ) ) _lowerCAmelCase : Tuple = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 67.3986 ) < 1E-2 assert abs(result_mean.item() - 0.0878 ) < 1E-3 def __magic_name__ ( self : int ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.full_loop(set_alpha_to_one=A_ , beta_start=0.01 ) _lowerCAmelCase : str = torch.sum(torch.abs(A_ ) ) _lowerCAmelCase : Optional[Any] = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 230.0399 ) < 1E-2 assert abs(result_mean.item() - 0.2995 ) < 1E-3 def __magic_name__ ( self : List[Any] ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.full_loop(set_alpha_to_one=A_ , beta_start=0.01 ) _lowerCAmelCase : str = torch.sum(torch.abs(A_ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 186.9482 ) < 1E-2 assert abs(result_mean.item() - 0.2434 ) < 1E-3
503
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract __UpperCAmelCase = logging.get_logger(__name__) def _snake_case ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" return [ int(1000 * (box[0] / width) ), int(1000 * (box[1] / height) ), int(1000 * (box[2] / width) ), int(1000 * (box[3] / height) ), ] def _snake_case ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> str: """simple docstring""" _lowerCAmelCase : str = tesseract_config if tesseract_config is not None else "" # apply OCR _lowerCAmelCase : List[str] = to_pil_image(SCREAMING_SNAKE_CASE ) _lowerCAmelCase , _lowerCAmelCase : List[str] = pil_image.size _lowerCAmelCase : Union[str, Any] = pytesseract.image_to_data(SCREAMING_SNAKE_CASE , lang=SCREAMING_SNAKE_CASE , output_type="dict" , config=SCREAMING_SNAKE_CASE ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates _lowerCAmelCase : Dict = [idx for idx, word in enumerate(SCREAMING_SNAKE_CASE ) if not word.strip()] _lowerCAmelCase : Any = [word for idx, word in enumerate(SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] _lowerCAmelCase : Dict = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] _lowerCAmelCase : Optional[Any] = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] _lowerCAmelCase : Optional[Any] = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] _lowerCAmelCase : Optional[int] = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format _lowerCAmelCase : Union[str, Any] = [] for x, y, w, h in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowerCAmelCase : Tuple = [x, y, x + w, y + h] actual_boxes.append(SCREAMING_SNAKE_CASE ) # finally, normalize the bounding boxes _lowerCAmelCase : Dict = [] for box in actual_boxes: normalized_boxes.append(normalize_box(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) assert len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ), "Not as many words as there are bounding boxes" return words, normalized_boxes class A__ ( A ): """simple docstring""" _lowercase : Union[str, Any] = ['''pixel_values'''] def __init__( self : Tuple , A_ : bool = True , A_ : Dict[str, int] = None , A_ : PILImageResampling = PILImageResampling.BILINEAR , A_ : bool = True , A_ : Optional[str] = None , A_ : Optional[str] = "" , **A_ : Dict , ): '''simple docstring''' super().__init__(**A_ ) _lowerCAmelCase : int = size if size is not None else {"height": 2_2_4, "width": 2_2_4} _lowerCAmelCase : Any = get_size_dict(A_ ) _lowerCAmelCase : Any = do_resize _lowerCAmelCase : Any = size _lowerCAmelCase : Dict = resample _lowerCAmelCase : Optional[int] = apply_ocr _lowerCAmelCase : Dict = ocr_lang _lowerCAmelCase : Optional[int] = tesseract_config def __magic_name__ ( self : Dict , A_ : np.ndarray , A_ : Dict[str, int] , A_ : PILImageResampling = PILImageResampling.BILINEAR , A_ : Optional[Union[str, ChannelDimension]] = None , **A_ : List[str] , ): '''simple docstring''' _lowerCAmelCase : Optional[int] = get_size_dict(A_ ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) _lowerCAmelCase : Dict = (size["height"], size["width"]) return resize(A_ , size=A_ , resample=A_ , data_format=A_ , **A_ ) def __magic_name__ ( self : Dict , A_ : ImageInput , A_ : bool = None , A_ : Dict[str, int] = None , A_ : PILImageResampling = None , A_ : bool = None , A_ : Optional[str] = None , A_ : Optional[str] = None , A_ : Optional[Union[str, TensorType]] = None , A_ : ChannelDimension = ChannelDimension.FIRST , **A_ : Dict , ): '''simple docstring''' _lowerCAmelCase : List[Any] = do_resize if do_resize is not None else self.do_resize _lowerCAmelCase : str = size if size is not None else self.size _lowerCAmelCase : List[Any] = get_size_dict(A_ ) _lowerCAmelCase : Tuple = resample if resample is not None else self.resample _lowerCAmelCase : Union[str, Any] = apply_ocr if apply_ocr is not None else self.apply_ocr _lowerCAmelCase : Tuple = ocr_lang if ocr_lang is not None else self.ocr_lang _lowerCAmelCase : str = tesseract_config if tesseract_config is not None else self.tesseract_config _lowerCAmelCase : Union[str, Any] = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) # All transformations expect numpy arrays. _lowerCAmelCase : Dict = [to_numpy_array(A_ ) for image in images] if apply_ocr: requires_backends(self , "pytesseract" ) _lowerCAmelCase : Any = [] _lowerCAmelCase : List[Any] = [] for image in images: _lowerCAmelCase , _lowerCAmelCase : Optional[int] = apply_tesseract(A_ , A_ , A_ ) words_batch.append(A_ ) boxes_batch.append(A_ ) if do_resize: _lowerCAmelCase : Union[str, Any] = [self.resize(image=A_ , size=A_ , resample=A_ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) _lowerCAmelCase : Optional[int] = [flip_channel_order(A_ ) for image in images] _lowerCAmelCase : List[Any] = [to_channel_dimension_format(A_ , A_ ) for image in images] _lowerCAmelCase : Tuple = BatchFeature(data={"pixel_values": images} , tensor_type=A_ ) if apply_ocr: _lowerCAmelCase : Tuple = words_batch _lowerCAmelCase : Any = boxes_batch return data
503
1
"""simple docstring""" from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( '''The RoBERTa Model transformer with early exiting (DeeRoBERTa). ''', A, ) class lowercase__ ( A ): '''simple docstring''' _UpperCAmelCase = RobertaConfig _UpperCAmelCase = '''roberta''' def __init__( self , snake_case ) -> int: super().__init__(__lowerCAmelCase ) _UpperCAmelCase = RobertaEmbeddings(__lowerCAmelCase ) self.init_weights() @add_start_docstrings( '''RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. ''', A, ) class lowercase__ ( A ): '''simple docstring''' _UpperCAmelCase = RobertaConfig _UpperCAmelCase = '''roberta''' def __init__( self , snake_case ) -> Dict: super().__init__(__lowerCAmelCase ) _UpperCAmelCase = config.num_labels _UpperCAmelCase = config.num_hidden_layers _UpperCAmelCase = DeeRobertaModel(__lowerCAmelCase ) _UpperCAmelCase = nn.Dropout(config.hidden_dropout_prob ) _UpperCAmelCase = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(__lowerCAmelCase ) def lowerCamelCase_ ( self , snake_case=None , snake_case=None , snake_case=None , snake_case=None , snake_case=None , snake_case=None , snake_case=None , snake_case=-1 , snake_case=False , ) -> Tuple: _UpperCAmelCase = self.num_layers try: _UpperCAmelCase = self.roberta( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , position_ids=__lowerCAmelCase , head_mask=__lowerCAmelCase , inputs_embeds=__lowerCAmelCase , ) _UpperCAmelCase = outputs[1] _UpperCAmelCase = self.dropout(__lowerCAmelCase ) _UpperCAmelCase = self.classifier(__lowerCAmelCase ) _UpperCAmelCase = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: _UpperCAmelCase = e.message _UpperCAmelCase = e.exit_layer _UpperCAmelCase = outputs[0] if not self.training: _UpperCAmelCase = entropy(__lowerCAmelCase ) _UpperCAmelCase = [] _UpperCAmelCase = [] if labels is not None: if self.num_labels == 1: # We are doing regression _UpperCAmelCase = MSELoss() _UpperCAmelCase = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: _UpperCAmelCase = CrossEntropyLoss() _UpperCAmelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits _UpperCAmelCase = [] for highway_exit in outputs[-1]: _UpperCAmelCase = highway_exit[0] if not self.training: highway_logits_all.append(__lowerCAmelCase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression _UpperCAmelCase = MSELoss() _UpperCAmelCase = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: _UpperCAmelCase = CrossEntropyLoss() _UpperCAmelCase = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(__lowerCAmelCase ) if train_highway: _UpperCAmelCase = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: _UpperCAmelCase = (loss,) + outputs if not self.training: _UpperCAmelCase = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: _UpperCAmelCase = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
573
import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets _a = "\n@inproceedings{xu-etal-2016-optimizing,\n title = {Optimizing Statistical Machine Translation for Text Simplification},\n authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris},\n journal = {Transactions of the Association for Computational Linguistics},\n volume = {4},\n year={2016},\n url = {https://www.aclweb.org/anthology/Q16-1029},\n pages = {401--415\n},\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" _a = "\\nWIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU\nIt can be used to evaluate the quality of machine-generated texts.\n" _a = "\nCalculates sari score (between 0 and 100) given a list of source and predicted\nsentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score.\nArgs:\n sources: list of source sentences where each sentence should be a string.\n predictions: list of predicted sentences where each sentence should be a string.\n references: list of lists of reference sentences where each sentence should be a string.\nReturns:\n sari: sari score\n sacrebleu: sacrebleu score\n exact: exact score\n\nExamples:\n >>> sources=[\"About 95 species are currently accepted .\"]\n >>> predictions=[\"About 95 you now get in .\"]\n >>> references=[[\"About 95 species are currently known .\"]]\n >>> wiki_split = datasets.load_metric(\"wiki_split\")\n >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references)\n >>> print(results)\n {'sari': 21.805555555555557, 'sacrebleu': 14.535768424205482, 'exact': 0.0}\n" def lowerCAmelCase__(__snake_case ) -> int: '''simple docstring''' def remove_articles(__snake_case ): lowerCamelCase__ = re.compile(R'''\b(a|an|the)\b''' ,re.UNICODE ) return re.sub(__snake_case ,''' ''' ,__snake_case ) def white_space_fix(__snake_case ): return " ".join(text.split() ) def remove_punc(__snake_case ): lowerCamelCase__ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__snake_case ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__snake_case ) ) ) ) def lowerCAmelCase__(__snake_case ,__snake_case ) -> Optional[int]: '''simple docstring''' return int(normalize_answer(__snake_case ) == normalize_answer(__snake_case ) ) def lowerCAmelCase__(__snake_case ,__snake_case ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase__ = [any(compute_exact(__snake_case ,__snake_case ) for ref in refs ) for pred, refs in zip(__snake_case ,__snake_case )] return (sum(__snake_case ) / len(__snake_case )) * 100 def lowerCAmelCase__(__snake_case ,__snake_case ,__snake_case ,__snake_case ) -> int: '''simple docstring''' lowerCamelCase__ = [rgram for rgrams in rgramslist for rgram in rgrams] lowerCamelCase__ = Counter(__snake_case ) lowerCamelCase__ = Counter(__snake_case ) lowerCamelCase__ = Counter() for sgram, scount in sgramcounter.items(): lowerCamelCase__ = scount * numref lowerCamelCase__ = Counter(__snake_case ) lowerCamelCase__ = Counter() for cgram, ccount in cgramcounter.items(): lowerCamelCase__ = ccount * numref # KEEP lowerCamelCase__ = sgramcounter_rep & cgramcounter_rep lowerCamelCase__ = keepgramcounter_rep & rgramcounter lowerCamelCase__ = sgramcounter_rep & rgramcounter lowerCamelCase__ = 0 lowerCamelCase__ = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. lowerCamelCase__ = 1 lowerCamelCase__ = 1 if len(__snake_case ) > 0: lowerCamelCase__ = keeptmpscorea / len(__snake_case ) if len(__snake_case ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) lowerCamelCase__ = keeptmpscorea / sum(keepgramcounterall_rep.values() ) lowerCamelCase__ = 0 if keepscore_precision > 0 or keepscore_recall > 0: lowerCamelCase__ = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION lowerCamelCase__ = sgramcounter_rep - cgramcounter_rep lowerCamelCase__ = delgramcounter_rep - rgramcounter lowerCamelCase__ = sgramcounter_rep - rgramcounter lowerCamelCase__ = 0 lowerCamelCase__ = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. lowerCamelCase__ = 1 if len(__snake_case ) > 0: lowerCamelCase__ = deltmpscorea / len(__snake_case ) # ADDITION lowerCamelCase__ = set(__snake_case ) - set(__snake_case ) lowerCamelCase__ = set(__snake_case ) & set(__snake_case ) lowerCamelCase__ = set(__snake_case ) - set(__snake_case ) lowerCamelCase__ = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. lowerCamelCase__ = 1 lowerCamelCase__ = 1 if len(__snake_case ) > 0: lowerCamelCase__ = addtmpscore / len(__snake_case ) if len(__snake_case ) > 0: lowerCamelCase__ = addtmpscore / len(__snake_case ) lowerCamelCase__ = 0 if addscore_precision > 0 or addscore_recall > 0: lowerCamelCase__ = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def lowerCAmelCase__(__snake_case ,__snake_case ,__snake_case ) -> List[str]: '''simple docstring''' lowerCamelCase__ = len(__snake_case ) lowerCamelCase__ = ssent.split(''' ''' ) lowerCamelCase__ = csent.split(''' ''' ) lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] for rsent in rsents: lowerCamelCase__ = rsent.split(''' ''' ) lowerCamelCase__ = [] lowerCamelCase__ = [] lowerCamelCase__ = [] ragramslist.append(__snake_case ) for i in range(0 ,len(__snake_case ) - 1 ): if i < len(__snake_case ) - 1: lowerCamelCase__ = ragrams[i] + ''' ''' + ragrams[i + 1] ragrams.append(__snake_case ) if i < len(__snake_case ) - 2: lowerCamelCase__ = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] ragrams.append(__snake_case ) if i < len(__snake_case ) - 3: lowerCamelCase__ = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] + ''' ''' + ragrams[i + 3] ragrams.append(__snake_case ) ragramslist.append(__snake_case ) ragramslist.append(__snake_case ) ragramslist.append(__snake_case ) for i in range(0 ,len(__snake_case ) - 1 ): if i < len(__snake_case ) - 1: lowerCamelCase__ = sagrams[i] + ''' ''' + sagrams[i + 1] sagrams.append(__snake_case ) if i < len(__snake_case ) - 2: lowerCamelCase__ = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] sagrams.append(__snake_case ) if i < len(__snake_case ) - 3: lowerCamelCase__ = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] + ''' ''' + sagrams[i + 3] sagrams.append(__snake_case ) for i in range(0 ,len(__snake_case ) - 1 ): if i < len(__snake_case ) - 1: lowerCamelCase__ = cagrams[i] + ''' ''' + cagrams[i + 1] cagrams.append(__snake_case ) if i < len(__snake_case ) - 2: lowerCamelCase__ = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] cagrams.append(__snake_case ) if i < len(__snake_case ) - 3: lowerCamelCase__ = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] + ''' ''' + cagrams[i + 3] cagrams.append(__snake_case ) ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) = SARIngram(__snake_case ,__snake_case ,__snake_case ,__snake_case ) ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) = SARIngram(__snake_case ,__snake_case ,__snake_case ,__snake_case ) ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) = SARIngram(__snake_case ,__snake_case ,__snake_case ,__snake_case ) ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) = SARIngram(__snake_case ,__snake_case ,__snake_case ,__snake_case ) lowerCamelCase__ = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 lowerCamelCase__ = sum([delascore, delascore, delascore, delascore] ) / 4 lowerCamelCase__ = sum([addascore, addascore, addascore, addascore] ) / 4 lowerCamelCase__ = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def lowerCAmelCase__(__snake_case ,__snake_case = True ,__snake_case = "13a" ,__snake_case = True ) -> Tuple: '''simple docstring''' if lowercase: lowerCamelCase__ = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: lowerCamelCase__ = sacrebleu.metrics.bleu._get_tokenizer(__snake_case )()(__snake_case ) else: lowerCamelCase__ = sacrebleu.TOKENIZERS[tokenizer]()(__snake_case ) elif tokenizer == "moses": lowerCamelCase__ = sacremoses.MosesTokenizer().tokenize(__snake_case ,return_str=__snake_case ,escape=__snake_case ) elif tokenizer == "penn": lowerCamelCase__ = sacremoses.MosesTokenizer().penn_tokenize(__snake_case ,return_str=__snake_case ) else: lowerCamelCase__ = sentence if not return_str: lowerCamelCase__ = normalized_sent.split() return normalized_sent def lowerCAmelCase__(__snake_case ,__snake_case ,__snake_case ) -> List[Any]: '''simple docstring''' if not (len(__snake_case ) == len(__snake_case ) == len(__snake_case )): raise ValueError('''Sources length must match predictions and references lengths.''' ) lowerCamelCase__ = 0 for src, pred, refs in zip(__snake_case ,__snake_case ,__snake_case ): sari_score += SARIsent(normalize(__snake_case ) ,normalize(__snake_case ) ,[normalize(__snake_case ) for sent in refs] ) lowerCamelCase__ = sari_score / len(__snake_case ) return 100 * sari_score def lowerCAmelCase__(__snake_case ,__snake_case ,__snake_case="exp" ,__snake_case=None ,__snake_case=False ,__snake_case=False ,__snake_case=False ,) -> int: '''simple docstring''' lowerCamelCase__ = len(references[0] ) if any(len(__snake_case ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) lowerCamelCase__ = [[refs[i] for refs in references] for i in range(__snake_case )] lowerCamelCase__ = sacrebleu.corpus_bleu( __snake_case ,__snake_case ,smooth_method=__snake_case ,smooth_value=__snake_case ,force=__snake_case ,lowercase=__snake_case ,use_effective_order=__snake_case ,) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): '''simple docstring''' def __lowerCamelCase ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=[ '''https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py''', '''https://github.com/cocoxu/simplification/blob/master/SARI.py''', '''https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py''', '''https://github.com/mjpost/sacreBLEU''', ] , reference_urls=[ '''https://www.aclweb.org/anthology/Q16-1029.pdf''', '''https://github.com/mjpost/sacreBLEU''', '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = {} result.update({'''sari''': compute_sari(sources=__lowerCAmelCase , predictions=__lowerCAmelCase , references=__lowerCAmelCase )} ) result.update({'''sacrebleu''': compute_sacrebleu(predictions=__lowerCAmelCase , references=__lowerCAmelCase )} ) result.update({'''exact''': compute_em(predictions=__lowerCAmelCase , references=__lowerCAmelCase )} ) return result
481
0
'''simple docstring''' import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) __snake_case =logging.getLogger(__name__) @dataclass(frozen=__lowercase ) class UpperCAmelCase_ : lowerCamelCase : str lowerCamelCase : str lowerCamelCase : Optional[str] = None lowerCamelCase : Optional[str] = None lowerCamelCase : Optional[str] = None @dataclass(frozen=__lowercase ) class UpperCAmelCase_ : lowerCamelCase : List[int] lowerCamelCase : Optional[List[int]] = None lowerCamelCase : Optional[List[int]] = None lowerCamelCase : Optional[Union[int, float]] = None lowerCamelCase : Optional[int] = None if is_torch_available(): import torch from torch.utils.data import Dataset class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : List[InputFeatures] def __init__( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : PreTrainedTokenizer , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : List[str]=False , UpperCAmelCase__ : bool = False , ) -> Any: lowerCAmelCase = hans_processors[task]() lowerCAmelCase = os.path.join( UpperCAmelCase__ , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(UpperCAmelCase__ ) , UpperCAmelCase__ , ) , ) lowerCAmelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase , lowerCAmelCase = label_list[2], label_list[1] lowerCAmelCase = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase = cached_features_file + '.lock' with FileLock(UpperCAmelCase__ ): if os.path.exists(UpperCAmelCase__ ) and not overwrite_cache: logger.info(F'''Loading features from cached file {cached_features_file}''' ) lowerCAmelCase = torch.load(UpperCAmelCase__ ) else: logger.info(F'''Creating features from dataset file at {data_dir}''' ) lowerCAmelCase = ( processor.get_dev_examples(UpperCAmelCase__ ) if evaluate else processor.get_train_examples(UpperCAmelCase__ ) ) logger.info('Training examples: %s' , len(UpperCAmelCase__ ) ) lowerCAmelCase = hans_convert_examples_to_features(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) logger.info('Saving features into cached file %s' , UpperCAmelCase__ ) torch.save(self.features , UpperCAmelCase__ ) def __len__( self : List[str] ) -> Dict: return len(self.features ) def __getitem__( self : Tuple , UpperCAmelCase__ : List[str] ) -> InputFeatures: return self.features[i] def __UpperCAmelCase ( self : str ) -> Tuple: return self.label_list if is_tf_available(): import tensorflow as tf class UpperCAmelCase_ : lowerCamelCase : List[InputFeatures] def __init__( self : Tuple , UpperCAmelCase__ : str , UpperCAmelCase__ : PreTrainedTokenizer , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] = 1_2_8 , UpperCAmelCase__ : List[Any]=False , UpperCAmelCase__ : bool = False , ) -> Tuple: lowerCAmelCase = hans_processors[task]() lowerCAmelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase , lowerCAmelCase = label_list[2], label_list[1] lowerCAmelCase = label_list lowerCAmelCase = processor.get_dev_examples(UpperCAmelCase__ ) if evaluate else processor.get_train_examples(UpperCAmelCase__ ) lowerCAmelCase = hans_convert_examples_to_features(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 1_0_0_0_0 == 0: logger.info('Writing example %d of %d' % (ex_index, len(UpperCAmelCase__ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCAmelCase = tf.data.Dataset.from_generator( UpperCAmelCase__ , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def __UpperCAmelCase ( self : Any ) -> int: return self.dataset def __len__( self : Optional[int] ) -> Tuple: return len(self.features ) def __getitem__( self : Optional[Any] , UpperCAmelCase__ : str ) -> InputFeatures: return self.features[i] def __UpperCAmelCase ( self : str ) -> str: return self.label_list class UpperCAmelCase_ ( __lowercase ): def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : int ) -> Union[str, Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase__ , 'heuristics_train_set.txt' ) ) , 'train' ) def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : Optional[Any] ) -> List[Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase__ , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def __UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: return ["contradiction", "entailment", "neutral"] def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any] ) -> Dict: lowerCAmelCase = [] for i, line in enumerate(UpperCAmelCase__ ): if i == 0: continue lowerCAmelCase = '%s-%s' % (set_type, line[0]) lowerCAmelCase = line[5] lowerCAmelCase = line[6] lowerCAmelCase = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCAmelCase = line[0] examples.append(InputExample(guid=UpperCAmelCase__ , text_a=UpperCAmelCase__ , text_b=UpperCAmelCase__ , label=UpperCAmelCase__ , pairID=UpperCAmelCase__ ) ) return examples def a_ ( lowerCamelCase : List[InputExample] , lowerCamelCase : List[str] , lowerCamelCase : int , lowerCamelCase : PreTrainedTokenizer , ): lowerCAmelCase = {label: i for i, label in enumerate(lowerCamelCase )} lowerCAmelCase = [] for ex_index, example in tqdm.tqdm(enumerate(lowerCamelCase ) , desc='convert examples to features' ): if ex_index % 10000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCAmelCase = tokenizer( example.text_a , example.text_b , add_special_tokens=lowerCamelCase , max_length=lowerCamelCase , padding='max_length' , truncation=lowerCamelCase , return_overflowing_tokens=lowerCamelCase , ) lowerCAmelCase = label_map[example.label] if example.label in label_map else 0 lowerCAmelCase = int(example.pairID ) features.append(InputFeatures(**lowerCamelCase , label=lowerCamelCase , pairID=lowerCamelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(f'''guid: {example}''' ) logger.info(f'''features: {features[i]}''' ) return features __snake_case ={ """hans""": 3, } __snake_case ={ """hans""": HansProcessor, }
513
'''simple docstring''' import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __snake_case =logging.get_logger(__name__) __snake_case ={"""vocab_file""": """spiece.model"""} __snake_case ={ """vocab_file""": { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model""", """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model""" ), } } __snake_case ={ """google/bigbird-roberta-base""": 4_096, """google/bigbird-roberta-large""": 4_096, """google/bigbird-base-trivia-itc""": 4_096, } class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : Optional[Any] = VOCAB_FILES_NAMES lowerCamelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] = ['''input_ids''', '''attention_mask'''] lowerCamelCase : List[int] = [] def __init__( self : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : Union[str, Any]="<s>" , UpperCAmelCase__ : str="</s>" , UpperCAmelCase__ : List[Any]="<pad>" , UpperCAmelCase__ : Any="[SEP]" , UpperCAmelCase__ : List[Any]="[MASK]" , UpperCAmelCase__ : Optional[int]="[CLS]" , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : Tuple , ) -> None: lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else bos_token lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else eos_token lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else unk_token lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else pad_token lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else cls_token lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) lowerCAmelCase = vocab_file lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) @property def __UpperCAmelCase ( self : List[str] ) -> Tuple: return self.sp_model.get_piece_size() def __UpperCAmelCase ( self : Dict ) -> Tuple: lowerCAmelCase = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ) -> Optional[int]: lowerCAmelCase = self.__dict__.copy() lowerCAmelCase = None return state def __setstate__( self : int , UpperCAmelCase__ : List[Any] ) -> Dict: lowerCAmelCase = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCAmelCase = {} lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: return self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) def __UpperCAmelCase ( self : Tuple , UpperCAmelCase__ : Optional[int] ) -> Dict: return self.sp_model.piece_to_id(UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : int ) -> Optional[int]: lowerCAmelCase = self.sp_model.IdToPiece(UpperCAmelCase__ ) return token def __UpperCAmelCase ( self : int , UpperCAmelCase__ : List[str] ) -> Any: lowerCAmelCase = [] lowerCAmelCase = '' lowerCAmelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(UpperCAmelCase__ ) + token lowerCAmelCase = True lowerCAmelCase = [] else: current_sub_tokens.append(UpperCAmelCase__ ) lowerCAmelCase = False out_string += self.sp_model.decode(UpperCAmelCase__ ) return out_string.strip() def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : bool = True , **UpperCAmelCase__ : int , ) -> str: lowerCAmelCase = kwargs.pop('use_source_tokenizer' , UpperCAmelCase__ ) lowerCAmelCase = self.convert_ids_to_tokens(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCAmelCase = [] lowerCAmelCase = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCAmelCase__ ) ) lowerCAmelCase = [] sub_texts.append(UpperCAmelCase__ ) else: current_sub_text.append(UpperCAmelCase__ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCAmelCase__ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: lowerCAmelCase = re.sub(R' (\[(MASK|SEP)\])' , R'\1' , ' '.join(UpperCAmelCase__ ) ) else: lowerCAmelCase = ''.join(UpperCAmelCase__ ) lowerCAmelCase = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCAmelCase = self.clean_up_tokenization(UpperCAmelCase__ ) return clean_text else: return text def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase = os.path.join( UpperCAmelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , 'wb' ) as fi: lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,) def __UpperCAmelCase ( self : int , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase__ )) + [1] return [1] + ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1] def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
513
1
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin __SCREAMING_SNAKE_CASE = random.Random() def SCREAMING_SNAKE_CASE__ ( lowerCAmelCase_ : Optional[int] ,lowerCAmelCase_ : Tuple=1.0 ,lowerCAmelCase_ : Dict=None ,lowerCAmelCase_ : Optional[int]=None ) -> Optional[Any]: """simple docstring""" if rng is None: SCREAMING_SNAKE_CASE_ : List[str] =global_rng SCREAMING_SNAKE_CASE_ : List[str] =[] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=7 , __UpperCAmelCase=400 , __UpperCAmelCase=2_000 , __UpperCAmelCase=1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=16_000 , __UpperCAmelCase=True , __UpperCAmelCase=True , ): SCREAMING_SNAKE_CASE_ : Optional[int] =parent SCREAMING_SNAKE_CASE_ : Any =batch_size SCREAMING_SNAKE_CASE_ : Optional[Any] =min_seq_length SCREAMING_SNAKE_CASE_ : Optional[int] =max_seq_length SCREAMING_SNAKE_CASE_ : Union[str, Any] =(self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) SCREAMING_SNAKE_CASE_ : Dict =feature_size SCREAMING_SNAKE_CASE_ : Dict =padding_value SCREAMING_SNAKE_CASE_ : List[Any] =sampling_rate SCREAMING_SNAKE_CASE_ : Any =return_attention_mask SCREAMING_SNAKE_CASE_ : Union[str, Any] =do_normalize def __lowerCamelCase ( self ): return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def __lowerCamelCase ( self , __UpperCAmelCase=False , __UpperCAmelCase=False ): def _flatten(__UpperCAmelCase ): return list(itertools.chain(*__UpperCAmelCase ) ) if equal_length: SCREAMING_SNAKE_CASE_ : List[str] =floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size SCREAMING_SNAKE_CASE_ : List[Any] =[ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: SCREAMING_SNAKE_CASE_ : str =[np.asarray(__UpperCAmelCase ) for x in speech_inputs] return speech_inputs class lowerCAmelCase_ ( __A , unittest.TestCase ): '''simple docstring''' _lowercase = WavaVecaFeatureExtractor def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : List[str] =WavaVecaFeatureExtractionTester(self ) def __lowerCamelCase ( self , __UpperCAmelCase ): self.assertTrue(np.all(np.mean(__UpperCAmelCase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__UpperCAmelCase , axis=0 ) - 1 ) < 1E-3 ) ) def __lowerCamelCase ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus SCREAMING_SNAKE_CASE_ : Union[str, Any] =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 SCREAMING_SNAKE_CASE_ : int =[floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] SCREAMING_SNAKE_CASE_ : Optional[Any] =[np.asarray(__UpperCAmelCase ) for speech_input in speech_inputs] # Test not batched input SCREAMING_SNAKE_CASE_ : List[Any] =feat_extract(speech_inputs[0] , return_tensors='np' ).input_values SCREAMING_SNAKE_CASE_ : List[str] =feat_extract(np_speech_inputs[0] , return_tensors='np' ).input_values self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) # Test batched SCREAMING_SNAKE_CASE_ : List[Any] =feat_extract(__UpperCAmelCase , return_tensors='np' ).input_values SCREAMING_SNAKE_CASE_ : Tuple =feat_extract(__UpperCAmelCase , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(__UpperCAmelCase , __UpperCAmelCase ): self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. SCREAMING_SNAKE_CASE_ : List[str] =[floats_list((1, x) )[0] for x in (800, 800, 800)] SCREAMING_SNAKE_CASE_ : str =np.asarray(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ : Dict =feat_extract(__UpperCAmelCase , return_tensors='np' ).input_values SCREAMING_SNAKE_CASE_ : List[Any] =feat_extract(__UpperCAmelCase , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(__UpperCAmelCase , __UpperCAmelCase ): self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-3 ) ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : List[str] =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE_ : str =[floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] SCREAMING_SNAKE_CASE_ : str =['longest', 'max_length', 'do_not_pad'] SCREAMING_SNAKE_CASE_ : Dict =[None, 1_600, None] for max_length, padding in zip(__UpperCAmelCase , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple =feat_extract(__UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , return_tensors='np' ) SCREAMING_SNAKE_CASE_ : int =processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self.assertTrue(input_values[0][1_000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Tuple =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE_ : int =range(800 , 1_400 , 200 ) SCREAMING_SNAKE_CASE_ : int =[floats_list((1, x) )[0] for x in lengths] SCREAMING_SNAKE_CASE_ : int =['longest', 'max_length', 'do_not_pad'] SCREAMING_SNAKE_CASE_ : List[Any] =[None, 1_600, None] for max_length, padding in zip(__UpperCAmelCase , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : int =feat_extract(__UpperCAmelCase , max_length=__UpperCAmelCase , padding=__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] =processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Union[str, Any] =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE_ : int =[floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] SCREAMING_SNAKE_CASE_ : Optional[Any] =feat_extract( __UpperCAmelCase , truncation=__UpperCAmelCase , max_length=1_000 , padding='max_length' , return_tensors='np' ) SCREAMING_SNAKE_CASE_ : str =processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : int =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =[floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] SCREAMING_SNAKE_CASE_ : int =feat_extract( __UpperCAmelCase , truncation=__UpperCAmelCase , max_length=1_000 , padding='longest' , return_tensors='np' ) SCREAMING_SNAKE_CASE_ : List[Any] =processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1_000) ) SCREAMING_SNAKE_CASE_ : List[str] =[floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] SCREAMING_SNAKE_CASE_ : List[Any] =feat_extract( __UpperCAmelCase , truncation=__UpperCAmelCase , max_length=2_000 , padding='longest' , return_tensors='np' ) SCREAMING_SNAKE_CASE_ : str =processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1_200) ) @require_torch def __lowerCamelCase ( self ): import torch SCREAMING_SNAKE_CASE_ : Tuple =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =np.random.rand(100 ).astype(np.floataa ) SCREAMING_SNAKE_CASE_ : str =np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: SCREAMING_SNAKE_CASE_ : Tuple =feature_extractor.pad([{'input_values': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) SCREAMING_SNAKE_CASE_ : List[str] =feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def __lowerCamelCase ( self ): # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: SCREAMING_SNAKE_CASE_ : List[str] =WavaVecaConfig.from_pretrained(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] =WavaVecaFeatureExtractor.from_pretrained(__UpperCAmelCase ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer' )
220
from __future__ import annotations class lowerCAmelCase_ : '''simple docstring''' def __init__( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Any =order # a_{0} ... a_{k} SCREAMING_SNAKE_CASE_ : List[str] =[1.0] + [0.0] * order # b_{0} ... b_{k} SCREAMING_SNAKE_CASE_ : List[str] =[1.0] + [0.0] * order # x[n-1] ... x[n-k] SCREAMING_SNAKE_CASE_ : List[Any] =[0.0] * self.order # y[n-1] ... y[n-k] SCREAMING_SNAKE_CASE_ : Any =[0.0] * self.order def __lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): if len(__UpperCAmelCase ) < self.order: SCREAMING_SNAKE_CASE_ : str =[1.0, *a_coeffs] if len(__UpperCAmelCase ) != self.order + 1: SCREAMING_SNAKE_CASE_ : List[Any] =( F"""Expected a_coeffs to have {self.order + 1} elements """ F"""for {self.order}-order filter, got {len(__UpperCAmelCase )}""" ) raise ValueError(__UpperCAmelCase ) if len(__UpperCAmelCase ) != self.order + 1: SCREAMING_SNAKE_CASE_ : List[str] =( F"""Expected b_coeffs to have {self.order + 1} elements """ F"""for {self.order}-order filter, got {len(__UpperCAmelCase )}""" ) raise ValueError(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] =a_coeffs SCREAMING_SNAKE_CASE_ : int =b_coeffs def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : str =0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =(result + self.b_coeffs[0] * sample) / self.a_coeffs[0] SCREAMING_SNAKE_CASE_ : Tuple =self.input_history[:-1] SCREAMING_SNAKE_CASE_ : List[Any] =self.output_history[:-1] SCREAMING_SNAKE_CASE_ : Any =sample SCREAMING_SNAKE_CASE_ : Dict =result return result
220
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer __UpperCAmelCase : Any = logging.get_logger(__name__) __UpperCAmelCase : Any = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __UpperCAmelCase : List[Any] = { "vocab_file": { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt", }, "tokenizer_file": { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json" ), "google/realm-orqa-nq-openqa": ( "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json" ), "google/realm-orqa-nq-reader": ( "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json" ), "google/realm-orqa-wq-openqa": ( "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json" ), "google/realm-orqa-wq-reader": ( "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json" ), }, } __UpperCAmelCase : List[Any] = { "google/realm-cc-news-pretrained-embedder": 512, "google/realm-cc-news-pretrained-encoder": 512, "google/realm-cc-news-pretrained-scorer": 512, "google/realm-cc-news-pretrained-openqa": 512, "google/realm-orqa-nq-openqa": 512, "google/realm-orqa-nq-reader": 512, "google/realm-orqa-wq-openqa": 512, "google/realm-orqa-wq-reader": 512, } __UpperCAmelCase : Any = { "google/realm-cc-news-pretrained-embedder": {"do_lower_case": True}, "google/realm-cc-news-pretrained-encoder": {"do_lower_case": True}, "google/realm-cc-news-pretrained-scorer": {"do_lower_case": True}, "google/realm-cc-news-pretrained-openqa": {"do_lower_case": True}, "google/realm-orqa-nq-openqa": {"do_lower_case": True}, "google/realm-orqa-nq-reader": {"do_lower_case": True}, "google/realm-orqa-wq-openqa": {"do_lower_case": True}, "google/realm-orqa-wq-reader": {"do_lower_case": True}, } class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = RealmTokenizer def __init__( self : int , A : Tuple=None , A : List[Any]=None , A : Tuple=True , A : Union[str, Any]="[UNK]" , A : Optional[Any]="[SEP]" , A : str="[PAD]" , A : List[str]="[CLS]" , A : List[Any]="[MASK]" , A : Optional[Any]=True , A : List[Any]=None , **A : Union[str, Any] , ): super().__init__( A , tokenizer_file=A , do_lower_case=A , unk_token=A , sep_token=A , pad_token=A , cls_token=A , mask_token=A , tokenize_chinese_chars=A , strip_accents=A , **A , ) __snake_case: int = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , A ) != do_lower_case or normalizer_state.get("""strip_accents""" , A ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , A ) != tokenize_chinese_chars ): __snake_case: List[Any] = getattr(A , normalizer_state.pop("""type""" ) ) __snake_case: List[str] = do_lower_case __snake_case: Optional[int] = strip_accents __snake_case: Union[str, Any] = tokenize_chinese_chars __snake_case: Optional[int] = normalizer_class(**A ) __snake_case: Any = do_lower_case def UpperCAmelCase__ ( self : Dict , A : List[Any] , **A : Union[str, Any] ): __snake_case: Dict = PaddingStrategy.MAX_LENGTH __snake_case: Tuple = text __snake_case: Optional[int] = kwargs.pop("""text_pair""" , A ) __snake_case: Optional[int] = kwargs.pop("""return_tensors""" , A ) __snake_case: Dict = { """input_ids""": [], """attention_mask""": [], """token_type_ids""": [], } for idx, candidate_text in enumerate(A ): if batch_text_pair is not None: __snake_case: Union[str, Any] = batch_text_pair[idx] else: __snake_case: Optional[Any] = None __snake_case: Union[str, Any] = super().__call__(A , A , return_tensors=A , **A ) __snake_case: str = encoded_candidates.get("""input_ids""" ) __snake_case: Dict = encoded_candidates.get("""attention_mask""" ) __snake_case: Dict = encoded_candidates.get("""token_type_ids""" ) if encoded_input_ids is not None: output_data["input_ids"].append(A ) if encoded_attention_mask is not None: output_data["attention_mask"].append(A ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(A ) __snake_case: int = {key: item for key, item in output_data.items() if len(A ) != 0} return BatchEncoding(A , tensor_type=A ) def UpperCAmelCase__ ( self : Union[str, Any] , A : Optional[int] , A : Dict=None ): __snake_case: Union[str, Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase__ ( self : List[Any] , A : List[int] , A : Optional[List[int]] = None ): __snake_case: Optional[Any] = [self.sep_token_id] __snake_case: List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self : List[Any] , A : str , A : Optional[str] = None ): __snake_case: Union[str, Any] = self._tokenizer.model.save(A , name=A ) return tuple(A )
706
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __snake_case ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' @register_to_config def __init__( self : int , A : int , A : int , A : int , A : float , A : int , A : int , A : int , A : int , A : str , A : bool = False , ): super().__init__() __snake_case: List[str] = nn.Embedding(A , A ) __snake_case: Optional[Any] = nn.Embedding(A , A ) __snake_case: Optional[Any] = False __snake_case: Optional[Any] = nn.Dropout(p=A ) __snake_case: Optional[Any] = TaConfig( vocab_size=A , d_model=A , num_heads=A , d_kv=A , d_ff=A , dropout_rate=A , feed_forward_proj=A , is_decoder=A , is_encoder_decoder=A , ) __snake_case: Union[str, Any] = nn.ModuleList() for lyr_num in range(A ): __snake_case: Optional[int] = TaBlock(A ) self.encoders.append(A ) __snake_case: Optional[int] = TaLayerNorm(A ) __snake_case: List[Any] = nn.Dropout(p=A ) def UpperCAmelCase__ ( self : Union[str, Any] , A : int , A : Any ): __snake_case: Tuple = self.token_embedder(A ) __snake_case: Union[str, Any] = encoder_input_tokens.shape[1] __snake_case: Dict = torch.arange(A , device=encoder_input_tokens.device ) x += self.position_encoding(A ) __snake_case: Optional[Any] = self.dropout_pre(A ) # inverted the attention mask __snake_case: Dict = encoder_input_tokens.size() __snake_case: str = self.get_extended_attention_mask(A , A ) for lyr in self.encoders: __snake_case: List[str] = lyr(A , A )[0] __snake_case: Any = self.layer_norm(A ) return self.dropout_post(A ), encoder_inputs_mask
155
0
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCAmelCase_ : int = '.' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) UpperCAmelCase_ : str = [ 'Assert', 'AssignVariableOp', 'EmptyTensorList', 'MergeV2Checkpoints', 'ReadVariableOp', 'ResourceGather', 'RestoreV2', 'SaveV2', 'ShardedFilename', 'StatefulPartitionedCall', 'StaticRegexFullMatch', 'VarHandleOp', ] def A_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : Any = SavedModel() _lowerCamelCase : Union[str, Any] = [] with open(os.path.join(_lowerCAmelCase , "utils" , "tf_ops" , "onnx.json" ) ) as f: _lowerCamelCase : List[Any] = json.load(_lowerCAmelCase )["opsets"] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(_lowerCAmelCase )] ) with open(_lowerCAmelCase , "rb" ) as f: saved_model.ParseFromString(f.read() ) _lowerCamelCase : Optional[Any] = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want _lowerCamelCase : List[str] = sorted(_lowerCAmelCase ) _lowerCamelCase : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(_lowerCAmelCase ) if strict and len(_lowerCAmelCase ) > 0: raise Exception(F'Found the following incompatible ops for the opset {opset}:\n' + incompatible_ops ) elif len(_lowerCAmelCase ) > 0: print(F'Found the following incompatible ops for the opset {opset}:' ) print(*_lowerCAmelCase , sep="\n" ) else: print(F'The saved model {saved_model_path} can properly be converted with ONNX.' ) if __name__ == "__main__": UpperCAmelCase_ : Any = argparse.ArgumentParser() parser.add_argument('--saved_model_path', help='Path of the saved model to check (the .pb file).') parser.add_argument( '--opset', default=12, type=int, help='The ONNX opset against which the model has to be tested.' ) parser.add_argument( '--framework', choices=['onnx'], default='onnx', help='Frameworks against which to test the saved model.' ) parser.add_argument( '--strict', action='store_true', help='Whether make the checking strict (raise errors) or not (raise warnings)' ) UpperCAmelCase_ : Union[str, Any] = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
44
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase_ = 1_6 lowercase_ = 3_2 def lowercase ( lowerCAmelCase__ : Accelerator , lowerCAmelCase__ : int = 16 , lowerCAmelCase__ : str = "bert-base-cased" ) -> Optional[int]: __a = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) __a = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(lowerCAmelCase__ : Optional[Any] ): # max_length=None => use the model max length (it's actually the default) __a = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __a = datasets.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=lowerCAmelCase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __a = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(lowerCAmelCase__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase__ , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(lowerCAmelCase__ , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __a = DataLoader( tokenized_datasets['''train'''] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) __a = DataLoader( tokenized_datasets['''validation'''] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) return train_dataloader, eval_dataloader def lowercase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Union[str, Any] ) -> Optional[int]: # Initialize accelerator __a = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __a = config['''lr'''] __a = int(config['''num_epochs'''] ) __a = int(config['''seed'''] ) __a = int(config['''batch_size'''] ) __a = args.model_name_or_path set_seed(lowerCAmelCase__ ) __a , __a = get_dataloaders(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __a = AutoModelForSequenceClassification.from_pretrained(lowerCAmelCase__ , return_dict=lowerCAmelCase__ ) # Instantiate optimizer __a = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __a = optimizer_cls(params=model.parameters() , lr=lowerCAmelCase__ ) if accelerator.state.deepspeed_plugin is not None: __a = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __a = 1 __a = (len(lowerCAmelCase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __a = get_linear_schedule_with_warmup( optimizer=lowerCAmelCase__ , num_warmup_steps=0 , num_training_steps=lowerCAmelCase__ , ) else: __a = DummyScheduler(lowerCAmelCase__ , total_num_steps=lowerCAmelCase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __a , __a , __a , __a , __a = accelerator.prepare( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # We need to keep track of how many total steps we have iterated over __a = 0 # We also need to keep track of the stating epoch so files are named properly __a = 0 # Now we train the model __a = evaluate.load('''glue''' , '''mrpc''' ) __a = 0 __a = {} for epoch in range(lowerCAmelCase__ , lowerCAmelCase__ ): model.train() for step, batch in enumerate(lowerCAmelCase__ ): __a = model(**lowerCAmelCase__ ) __a = outputs.loss __a = loss / gradient_accumulation_steps accelerator.backward(lowerCAmelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() __a = 0 for step, batch in enumerate(lowerCAmelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __a = model(**lowerCAmelCase__ ) __a = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __a , __a = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowerCAmelCase__ ) - 1: __a = predictions[: len(eval_dataloader.dataset ) - samples_seen] __a = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowerCAmelCase__ , references=lowerCAmelCase__ , ) __a = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowerCAmelCase__ ) __a = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: __a = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) def lowercase ( ) -> List[str]: __a = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=lowerCAmelCase__ , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=lowerCAmelCase__ , ) parser.add_argument( '''--output_dir''' , type=lowerCAmelCase__ , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=lowerCAmelCase__ , default=3 , help='''Number of train epochs.''' , ) __a = parser.parse_args() __a = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(lowerCAmelCase__ , lowerCAmelCase__ ) if __name__ == "__main__": main()
695
0
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def _lowerCAmelCase ( _lowerCAmelCase = 8 ): '''simple docstring''' A_ : List[str] = ascii_letters + digits + punctuation return "".join(secrets.choice(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ) ) def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ): '''simple docstring''' i -= len(_lowerCAmelCase ) A_ : List[str] = i // 3 A_ : Tuple = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) A_ : int = ( chars_incl + random(_lowerCAmelCase ,quotient + remainder ) + random(_lowerCAmelCase ,_lowerCAmelCase ) + random(_lowerCAmelCase ,_lowerCAmelCase ) ) A_ : List[Any] = list(_lowerCAmelCase ) shuffle(_lowerCAmelCase ) return "".join(_lowerCAmelCase ) # random is a generalised function for letters, characters and numbers def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ): '''simple docstring''' return "".join(secrets.choice(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ) ) def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ): '''simple docstring''' pass # Put your code here... def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ): '''simple docstring''' pass # Put your code here... def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ): '''simple docstring''' pass # Put your code here... def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase = 8 ): '''simple docstring''' if len(_lowerCAmelCase ) < min_length: # Your Password must be at least 8 characters long return False A_ : List[Any] = any(char in ascii_uppercase for char in password ) A_ : int = any(char in ascii_lowercase for char in password ) A_ : Dict = any(char in digits for char in password ) A_ : int = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def _lowerCAmelCase ( ): '''simple docstring''' A_ : int = int(input("""Please indicate the max length of your password: """ ).strip() ) A_ : Tuple = input( """Please indicate the characters that must be in your password: """ ).strip() print("""Password generated:""" ,password_generator(_lowerCAmelCase ) ) print( """Alternative Password generated:""" ,alternative_password_generator(_lowerCAmelCase ,_lowerCAmelCase ) ,) print("""[If you are thinking of using this passsword, You better save it.]""" ) if __name__ == "__main__": main()
713
def _lowerCAmelCase ( _lowerCAmelCase ): '''simple docstring''' A_ : Union[str, Any] = [0] * len(_lowerCAmelCase ) A_ : Optional[int] = [] A_ : str = [] A_ : Dict = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(_lowerCAmelCase ) ): if indegree[i] == 0: queue.append(_lowerCAmelCase ) while queue: A_ : List[str] = queue.pop(0 ) cnt += 1 topo.append(_lowerCAmelCase ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(_lowerCAmelCase ) if cnt != len(_lowerCAmelCase ): print("""Cycle exists""" ) else: print(_lowerCAmelCase ) # Adjacency List of Graph _lowerCAmelCase = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
481
0
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class a_ ( __UpperCamelCase ): UpperCamelCase_ : "DiagonalGaussianDistribution" class a_ ( __UpperCamelCase , __UpperCamelCase ): UpperCamelCase_ : Union[str, Any] = True @register_to_config def __init__( self : int , snake_case__ : int = 3 , snake_case__ : int = 3 , snake_case__ : Tuple[str] = ("DownEncoderBlock2D",) , snake_case__ : Tuple[str] = ("UpDecoderBlock2D",) , snake_case__ : Tuple[int] = (64,) , snake_case__ : int = 1 , snake_case__ : str = "silu" , snake_case__ : int = 4 , snake_case__ : int = 32 , snake_case__ : int = 32 , snake_case__ : float = 0.1_8215 , ): super().__init__() # pass init params to Encoder lowerCAmelCase__ = Encoder( in_channels=snake_case__ , out_channels=snake_case__ , down_block_types=snake_case__ , block_out_channels=snake_case__ , layers_per_block=snake_case__ , act_fn=snake_case__ , norm_num_groups=snake_case__ , double_z=snake_case__ , ) # pass init params to Decoder lowerCAmelCase__ = Decoder( in_channels=snake_case__ , out_channels=snake_case__ , up_block_types=snake_case__ , block_out_channels=snake_case__ , layers_per_block=snake_case__ , norm_num_groups=snake_case__ , act_fn=snake_case__ , ) lowerCAmelCase__ = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) lowerCAmelCase__ = nn.Convad(snake_case__ , snake_case__ , 1 ) lowerCAmelCase__ = False lowerCAmelCase__ = False # only relevant if vae tiling is enabled lowerCAmelCase__ = self.config.sample_size lowerCAmelCase__ = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) lowerCAmelCase__ = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) lowerCAmelCase__ = 0.25 def _SCREAMING_SNAKE_CASE ( self : str , snake_case__ : int , snake_case__ : Tuple=False ): if isinstance(snake_case__ , (Encoder, Decoder) ): lowerCAmelCase__ = value def _SCREAMING_SNAKE_CASE ( self : Dict , snake_case__ : bool = True ): lowerCAmelCase__ = use_tiling def _SCREAMING_SNAKE_CASE ( self : List[Any] ): self.enable_tiling(snake_case__ ) def _SCREAMING_SNAKE_CASE ( self : str ): lowerCAmelCase__ = True def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): lowerCAmelCase__ = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def _SCREAMING_SNAKE_CASE ( self : List[Any] ): lowerCAmelCase__ = {} def fn_recursive_add_processors(snake_case__ : str , snake_case__ : torch.nn.Module , snake_case__ : Dict[str, AttentionProcessor] ): if hasattr(snake_case__ , """set_processor""" ): lowerCAmelCase__ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"""{name}.{sub_name}""" , snake_case__ , snake_case__ ) return processors for name, module in self.named_children(): fn_recursive_add_processors(snake_case__ , snake_case__ , snake_case__ ) return processors def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , snake_case__ : Union[AttentionProcessor, Dict[str, AttentionProcessor]] ): lowerCAmelCase__ = len(self.attn_processors.keys() ) if isinstance(snake_case__ , snake_case__ ) and len(snake_case__ ) != count: raise ValueError( F"""A dict of processors was passed, but the number of processors {len(snake_case__ )} does not match the""" F""" number of attention layers: {count}. Please make sure to pass {count} processor classes.""" ) def fn_recursive_attn_processor(snake_case__ : str , snake_case__ : torch.nn.Module , snake_case__ : Union[str, Any] ): if hasattr(snake_case__ , """set_processor""" ): if not isinstance(snake_case__ , snake_case__ ): module.set_processor(snake_case__ ) else: module.set_processor(processor.pop(F"""{name}.processor""" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"""{name}.{sub_name}""" , snake_case__ , snake_case__ ) for name, module in self.named_children(): fn_recursive_attn_processor(snake_case__ , snake_case__ , snake_case__ ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def _SCREAMING_SNAKE_CASE ( self : Optional[int] , snake_case__ : torch.FloatTensor , snake_case__ : bool = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(snake_case__ , return_dict=snake_case__ ) if self.use_slicing and x.shape[0] > 1: lowerCAmelCase__ = [self.encoder(snake_case__ ) for x_slice in x.split(1 )] lowerCAmelCase__ = torch.cat(snake_case__ ) else: lowerCAmelCase__ = self.encoder(snake_case__ ) lowerCAmelCase__ = self.quant_conv(snake_case__ ) lowerCAmelCase__ = DiagonalGaussianDistribution(snake_case__ ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=snake_case__ ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , snake_case__ : torch.FloatTensor , snake_case__ : bool = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(snake_case__ , return_dict=snake_case__ ) lowerCAmelCase__ = self.post_quant_conv(snake_case__ ) lowerCAmelCase__ = self.decoder(snake_case__ ) if not return_dict: return (dec,) return DecoderOutput(sample=snake_case__ ) @apply_forward_hook def _SCREAMING_SNAKE_CASE ( self : Any , snake_case__ : torch.FloatTensor , snake_case__ : bool = True ): if self.use_slicing and z.shape[0] > 1: lowerCAmelCase__ = [self._decode(snake_case__ ).sample for z_slice in z.split(1 )] lowerCAmelCase__ = torch.cat(snake_case__ ) else: lowerCAmelCase__ = self._decode(snake_case__ ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=snake_case__ ) def _SCREAMING_SNAKE_CASE ( self : Any , snake_case__ : int , snake_case__ : Optional[int] , snake_case__ : str ): lowerCAmelCase__ = min(a.shape[2] , b.shape[2] , snake_case__ ) for y in range(snake_case__ ): lowerCAmelCase__ = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def _SCREAMING_SNAKE_CASE ( self : Optional[int] , snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : Optional[int] ): lowerCAmelCase__ = min(a.shape[3] , b.shape[3] , snake_case__ ) for x in range(snake_case__ ): lowerCAmelCase__ = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def _SCREAMING_SNAKE_CASE ( self : List[str] , snake_case__ : torch.FloatTensor , snake_case__ : bool = True ): lowerCAmelCase__ = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) lowerCAmelCase__ = int(self.tile_latent_min_size * self.tile_overlap_factor ) lowerCAmelCase__ = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. lowerCAmelCase__ = [] for i in range(0 , x.shape[2] , snake_case__ ): lowerCAmelCase__ = [] for j in range(0 , x.shape[3] , snake_case__ ): lowerCAmelCase__ = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] lowerCAmelCase__ = self.encoder(snake_case__ ) lowerCAmelCase__ = self.quant_conv(snake_case__ ) row.append(snake_case__ ) rows.append(snake_case__ ) lowerCAmelCase__ = [] for i, row in enumerate(snake_case__ ): lowerCAmelCase__ = [] for j, tile in enumerate(snake_case__ ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: lowerCAmelCase__ = self.blend_v(rows[i - 1][j] , snake_case__ , snake_case__ ) if j > 0: lowerCAmelCase__ = self.blend_h(row[j - 1] , snake_case__ , snake_case__ ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(snake_case__ , dim=3 ) ) lowerCAmelCase__ = torch.cat(snake_case__ , dim=2 ) lowerCAmelCase__ = DiagonalGaussianDistribution(snake_case__ ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=snake_case__ ) def _SCREAMING_SNAKE_CASE ( self : List[str] , snake_case__ : torch.FloatTensor , snake_case__ : bool = True ): lowerCAmelCase__ = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) lowerCAmelCase__ = int(self.tile_sample_min_size * self.tile_overlap_factor ) lowerCAmelCase__ = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. lowerCAmelCase__ = [] for i in range(0 , z.shape[2] , snake_case__ ): lowerCAmelCase__ = [] for j in range(0 , z.shape[3] , snake_case__ ): lowerCAmelCase__ = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] lowerCAmelCase__ = self.post_quant_conv(snake_case__ ) lowerCAmelCase__ = self.decoder(snake_case__ ) row.append(snake_case__ ) rows.append(snake_case__ ) lowerCAmelCase__ = [] for i, row in enumerate(snake_case__ ): lowerCAmelCase__ = [] for j, tile in enumerate(snake_case__ ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: lowerCAmelCase__ = self.blend_v(rows[i - 1][j] , snake_case__ , snake_case__ ) if j > 0: lowerCAmelCase__ = self.blend_h(row[j - 1] , snake_case__ , snake_case__ ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(snake_case__ , dim=3 ) ) lowerCAmelCase__ = torch.cat(snake_case__ , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=snake_case__ ) def _SCREAMING_SNAKE_CASE ( self : Tuple , snake_case__ : torch.FloatTensor , snake_case__ : bool = False , snake_case__ : bool = True , snake_case__ : Optional[torch.Generator] = None , ): lowerCAmelCase__ = sample lowerCAmelCase__ = self.encode(snake_case__ ).latent_dist if sample_posterior: lowerCAmelCase__ = posterior.sample(generator=snake_case__ ) else: lowerCAmelCase__ = posterior.mode() lowerCAmelCase__ = self.decode(snake_case__ ).sample if not return_dict: return (dec,) return DecoderOutput(sample=snake_case__ )
644
"""simple docstring""" def _UpperCAmelCase ( lowerCamelCase__ ): """simple docstring""" if len(lowerCamelCase__ ) <= 1: return lst lowerCAmelCase__ = 1 while i < len(lowerCamelCase__ ): if lst[i - 1] <= lst[i]: i += 1 else: lowerCAmelCase__ , lowerCAmelCase__ = lst[i], lst[i - 1] i -= 1 if i == 0: lowerCAmelCase__ = 1 return lst if __name__ == "__main__": __lowerCAmelCase : List[Any] = input("Enter numbers separated by a comma:\n").strip() __lowerCAmelCase : List[str] = [int(item) for item in user_input.split(",")] print(gnome_sort(unsorted))
644
1
"""simple docstring""" SCREAMING_SNAKE_CASE_ = { """A""": """.-""", """B""": """-...""", """C""": """-.-.""", """D""": """-..""", """E""": """.""", """F""": """..-.""", """G""": """--.""", """H""": """....""", """I""": """..""", """J""": """.---""", """K""": """-.-""", """L""": """.-..""", """M""": """--""", """N""": """-.""", """O""": """---""", """P""": """.--.""", """Q""": """--.-""", """R""": """.-.""", """S""": """...""", """T""": """-""", """U""": """..-""", """V""": """...-""", """W""": """.--""", """X""": """-..-""", """Y""": """-.--""", """Z""": """--..""", """1""": """.----""", """2""": """..---""", """3""": """...--""", """4""": """....-""", """5""": """.....""", """6""": """-....""", """7""": """--...""", """8""": """---..""", """9""": """----.""", """0""": """-----""", """&""": """.-...""", """@""": """.--.-.""", """:""": """---...""", """,""": """--..--""", """.""": """.-.-.-""", """\'""": """.----.""", """\"""": """.-..-.""", """?""": """..--..""", """/""": """-..-.""", """=""": """-...-""", """+""": """.-.-.""", """-""": """-....-""", """(""": """-.--.""", """)""": """-.--.-""", """!""": """-.-.--""", """ """: """/""" } # Exclamation mark is not in ITU-R recommendation # fmt: on SCREAMING_SNAKE_CASE_ = {value: key for key, value in MORSE_CODE_DICT.items()} def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ) -> str: return "".join(REVERSE_DICT[char] for char in message.split() ) def lowerCAmelCase_ ( ) -> None: a_ : Dict = "Morse code here!" print(snake_case__ ) a_ : List[str] = encrypt(snake_case__ ) print(snake_case__ ) a_ : Optional[int] = decrypt(snake_case__ ) print(snake_case__ ) if __name__ == "__main__": main()
717
"""simple docstring""" def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3_317_044_064_679_887_385_961_981 and not allow_probable: raise ValueError( "Warning: upper bound of deterministic test is exceeded. " "Pass allow_probable=True to allow probabilistic test. " "A return value of True indicates a probable prime." ) # array bounds provided by analysis a_ : Optional[Any] = [ 2_047, 1_373_653, 25_326_001, 3_215_031_751, 2_152_302_898_747, 3_474_749_660_383, 341_550_071_728_321, 1, 3_825_123_056_546_413_051, 1, 1, 318_665_857_834_031_151_167_461, 3_317_044_064_679_887_385_961_981, ] a_ : Optional[int] = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(SCREAMING_SNAKE_CASE__, 1 ): if n < _p: # then we have our last prime to check a_ : List[Any] = primes[:idx] break a_ , a_ : Tuple = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: a_ : str = False for r in range(SCREAMING_SNAKE_CASE__ ): a_ : List[str] = pow(SCREAMING_SNAKE_CASE__, d * 2**r, SCREAMING_SNAKE_CASE__ ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): a_ : List[Any] = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def lowerCAmelCase_ ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838_201 ) assert miller_rabin(838_207 ) # 1_373_653 assert not miller_rabin(17_316_001 ) assert miller_rabin(17_316_017 ) # 25_326_001 assert not miller_rabin(3_078_386_641 ) assert miller_rabin(3_078_386_653 ) # 3_215_031_751 assert not miller_rabin(1_713_045_574_801 ) assert miller_rabin(1_713_045_574_819 ) # 2_152_302_898_747 assert not miller_rabin(2_779_799_728_307 ) assert miller_rabin(2_779_799_728_327 ) # 3_474_749_660_383 assert not miller_rabin(113_850_023_909_441 ) assert miller_rabin(113_850_023_909_527 ) # 341_550_071_728_321 assert not miller_rabin(1_275_041_018_848_804_351 ) assert miller_rabin(1_275_041_018_848_804_391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79_666_464_458_507_787_791_867 ) assert miller_rabin(79_666_464_458_507_787_791_951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552_840_677_446_647_897_660_333 ) assert miller_rabin(552_840_677_446_647_897_660_359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
370
0
import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) _lowerCamelCase : Optional[int] = logging.getLogger() def a_ ( __lowercase : str ) -> Dict: _snake_case = {} _snake_case = os.path.join(__lowercase , 'all_results.json' ) if os.path.exists(__lowercase ): with open(__lowercase , 'r' ) as f: _snake_case = json.load(__lowercase ) else: raise ValueError(f'''can\'t find {path}''' ) return results _lowerCamelCase : Dict = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def A ( self : str ): '''simple docstring''' import xla_spawn _snake_case = self.get_auto_remove_tmp_dir() _snake_case = f''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowercase , 'argv' , lowercase ): _snake_case = time() xla_spawn.main() _snake_case = time() _snake_case = get_results(lowercase ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def A ( self : Union[str, Any] ): '''simple docstring''' import xla_spawn _snake_case = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowercase , 'argv' , lowercase ): xla_spawn.main()
686
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : Optional[Any] = logging.get_logger(__name__) def a_ ( __lowercase : Union[str, Any] ) -> List[Any]: _snake_case = SwinConfig( embed_dim=192 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=['stage2', 'stage3', 'stage4'] , ) _snake_case = DetaConfig( backbone_config=__lowercase , num_queries=900 , encoder_ffn_dim=2_048 , decoder_ffn_dim=2_048 , num_feature_levels=5 , assign_first_stage=__lowercase , with_box_refine=__lowercase , two_stage=__lowercase , ) # set labels _snake_case = 'huggingface/label-files' if "o365" in model_name: _snake_case = 366 _snake_case = 'object365-id2label.json' else: _snake_case = 91 _snake_case = 'coco-detection-id2label.json' _snake_case = num_labels _snake_case = json.load(open(cached_download(hf_hub_url(__lowercase , __lowercase , repo_type='dataset' ) ) , 'r' ) ) _snake_case = {int(__lowercase ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} return config def a_ ( __lowercase : int ) -> str: _snake_case = [] # stem # fmt: off rename_keys.append(('backbone.0.body.patch_embed.proj.weight', 'model.backbone.model.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.0.body.patch_embed.proj.bias', 'model.backbone.model.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.0.body.patch_embed.norm.weight', 'model.backbone.model.embeddings.norm.weight') ) rename_keys.append(('backbone.0.body.patch_embed.norm.bias', 'model.backbone.model.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.norm1.weight''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.norm1.bias''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.norm2.weight''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.norm2.bias''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias''', f'''model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((f'''backbone.0.body.layers.{i}.downsample.reduction.weight''', f'''model.backbone.model.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.downsample.norm.weight''', f'''model.backbone.model.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((f'''backbone.0.body.layers.{i}.downsample.norm.bias''', f'''model.backbone.model.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append(('backbone.0.body.norm1.weight', 'model.backbone.model.hidden_states_norms.stage2.weight') ) rename_keys.append(('backbone.0.body.norm1.bias', 'model.backbone.model.hidden_states_norms.stage2.bias') ) rename_keys.append(('backbone.0.body.norm2.weight', 'model.backbone.model.hidden_states_norms.stage3.weight') ) rename_keys.append(('backbone.0.body.norm2.bias', 'model.backbone.model.hidden_states_norms.stage3.bias') ) rename_keys.append(('backbone.0.body.norm3.weight', 'model.backbone.model.hidden_states_norms.stage4.weight') ) rename_keys.append(('backbone.0.body.norm3.bias', 'model.backbone.model.hidden_states_norms.stage4.bias') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight''', f'''model.encoder.layers.{i}.self_attn.sampling_offsets.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias''', f'''model.encoder.layers.{i}.self_attn.sampling_offsets.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.attention_weights.weight''', f'''model.encoder.layers.{i}.self_attn.attention_weights.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.attention_weights.bias''', f'''model.encoder.layers.{i}.self_attn.attention_weights.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.value_proj.weight''', f'''model.encoder.layers.{i}.self_attn.value_proj.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.value_proj.bias''', f'''model.encoder.layers.{i}.self_attn.value_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.output_proj.weight''', f'''model.encoder.layers.{i}.self_attn.output_proj.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.self_attn.output_proj.bias''', f'''model.encoder.layers.{i}.self_attn.output_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.weight''', f'''model.encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''model.encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''model.encoder.layers.{i}.fc1.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''model.encoder.layers.{i}.fc1.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''model.encoder.layers.{i}.fc2.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''model.encoder.layers.{i}.fc2.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''model.encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''model.encoder.layers.{i}.final_layer_norm.bias''') ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight''', f'''model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias''', f'''model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.attention_weights.weight''', f'''model.decoder.layers.{i}.encoder_attn.attention_weights.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.attention_weights.bias''', f'''model.decoder.layers.{i}.encoder_attn.attention_weights.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.value_proj.weight''', f'''model.decoder.layers.{i}.encoder_attn.value_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.value_proj.bias''', f'''model.decoder.layers.{i}.encoder_attn.value_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.output_proj.weight''', f'''model.decoder.layers.{i}.encoder_attn.output_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.cross_attn.output_proj.bias''', f'''model.decoder.layers.{i}.encoder_attn.output_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.weight''', f'''model.decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''model.decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''model.decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''model.decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm2.weight''', f'''model.decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm2.bias''', f'''model.decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''model.decoder.layers.{i}.fc1.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''model.decoder.layers.{i}.fc1.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''model.decoder.layers.{i}.fc2.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''model.decoder.layers.{i}.fc2.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''model.decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''model.decoder.layers.{i}.final_layer_norm.bias''') ) # fmt: on return rename_keys def a_ ( __lowercase : str , __lowercase : Tuple , __lowercase : str ) -> Union[str, Any]: _snake_case = dct.pop(__lowercase ) _snake_case = val def a_ ( __lowercase : List[str] , __lowercase : str ) -> Dict: _snake_case = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): _snake_case = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) _snake_case = state_dict.pop(f'''backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _snake_case = in_proj_weight[:dim, :] _snake_case = in_proj_bias[: dim] _snake_case = in_proj_weight[ dim : dim * 2, : ] _snake_case = in_proj_bias[ dim : dim * 2 ] _snake_case = in_proj_weight[ -dim :, : ] _snake_case = in_proj_bias[-dim :] # fmt: on def a_ ( __lowercase : Dict , __lowercase : Dict ) -> str: # transformer decoder self-attention layers _snake_case = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention _snake_case = state_dict.pop(f'''transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) _snake_case = state_dict.pop(f'''transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _snake_case = in_proj_weight[:hidden_size, :] _snake_case = in_proj_bias[:hidden_size] _snake_case = in_proj_weight[ hidden_size : hidden_size * 2, : ] _snake_case = in_proj_bias[hidden_size : hidden_size * 2] _snake_case = in_proj_weight[-hidden_size:, :] _snake_case = in_proj_bias[-hidden_size:] def a_ ( ) -> List[str]: _snake_case = 'http://images.cocodataset.org/val2017/000000039769.jpg' _snake_case = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ) return im @torch.no_grad() def a_ ( __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : Tuple ) -> Optional[Any]: _snake_case = get_deta_config(__lowercase ) # load original state dict if model_name == "deta-swin-large": _snake_case = hf_hub_download(repo_id='nielsr/deta-checkpoints' , filename='adet_swin_ft.pth' ) elif model_name == "deta-swin-large-o365": _snake_case = hf_hub_download(repo_id='jozhang97/deta-swin-l-o365' , filename='deta_swin_pt_o365.pth' ) else: raise ValueError(f'''Model name {model_name} not supported''' ) _snake_case = torch.load(__lowercase , map_location='cpu' )['model'] # original state dict for name, param in state_dict.items(): print(__lowercase , param.shape ) # rename keys _snake_case = create_rename_keys(__lowercase ) for src, dest in rename_keys: rename_key(__lowercase , __lowercase , __lowercase ) read_in_swin_q_k_v(__lowercase , config.backbone_config ) read_in_decoder_q_k_v(__lowercase , __lowercase ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: _snake_case = state_dict.pop(__lowercase ) _snake_case = val if "input_proj" in key: _snake_case = state_dict.pop(__lowercase ) _snake_case = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: _snake_case = state_dict.pop(__lowercase ) _snake_case = val # finally, create HuggingFace model and load state dict _snake_case = DetaForObjectDetection(__lowercase ) model.load_state_dict(__lowercase ) model.eval() _snake_case = 'cuda' if torch.cuda.is_available() else 'cpu' model.to(__lowercase ) # load image processor _snake_case = DetaImageProcessor(format='coco_detection' ) # verify our conversion on image _snake_case = prepare_img() _snake_case = processor(images=__lowercase , return_tensors='pt' ) _snake_case = encoding['pixel_values'] _snake_case = model(pixel_values.to(__lowercase ) ) # verify logits print('Logits:' , outputs.logits[0, :3, :3] ) print('Boxes:' , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": _snake_case = torch.tensor( [[-7.6_3_0_8, -2.8_4_8_5, -5.3_7_3_7], [-7.2_0_3_7, -4.5_5_0_5, -4.8_0_2_7], [-7.2_9_4_3, -4.2_6_1_1, -4.6_6_1_7]] ) _snake_case = torch.tensor([[0.4_9_8_7, 0.4_9_6_9, 0.9_9_9_9], [0.2_5_4_9, 0.5_4_9_8, 0.4_8_0_5], [0.5_4_9_8, 0.2_7_5_7, 0.0_5_6_9]] ) elif model_name == "deta-swin-large-o365": _snake_case = torch.tensor( [[-8.0_1_2_2, -3.5_7_2_0, -4.9_7_1_7], [-8.1_5_4_7, -3.6_8_8_6, -4.6_3_8_9], [-7.6_6_1_0, -3.6_1_9_4, -5.0_1_3_4]] ) _snake_case = torch.tensor([[0.2_5_2_3, 0.5_5_4_9, 0.4_8_8_1], [0.7_7_1_5, 0.4_1_4_9, 0.4_6_0_1], [0.5_5_0_3, 0.2_7_5_3, 0.0_5_7_5]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(__lowercase ) , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(__lowercase ) , atol=1E-4 ) print('Everything ok!' ) if pytorch_dump_folder_path: # Save model and processor logger.info(f'''Saving PyTorch model and processor to {pytorch_dump_folder_path}...''' ) Path(__lowercase ).mkdir(exist_ok=__lowercase ) model.save_pretrained(__lowercase ) processor.save_pretrained(__lowercase ) # Push to hub if push_to_hub: print('Pushing model and processor to hub...' ) model.push_to_hub(f'''jozhang97/{model_name}''' ) processor.push_to_hub(f'''jozhang97/{model_name}''' ) if __name__ == "__main__": _lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument( '''--model_name''', type=str, default='''deta-swin-large''', choices=['''deta-swin-large''', '''deta-swin-large-o365'''], help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _lowerCamelCase : List[Any] = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
686
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : Dict = { """microsoft/swinv2-tiny-patch4-window8-256""": ( """https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json""" ), } class lowerCAmelCase ( __UpperCAmelCase ): a : Dict = """swinv2""" a : Any = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , UpperCamelCase=224 , UpperCamelCase=4 , UpperCamelCase=3 , UpperCamelCase=96 , UpperCamelCase=[2, 2, 6, 2] , UpperCamelCase=[3, 6, 12, 24] , UpperCamelCase=7 , UpperCamelCase=4.0 , UpperCamelCase=True , UpperCamelCase=0.0 , UpperCamelCase=0.0 , UpperCamelCase=0.1 , UpperCamelCase="gelu" , UpperCamelCase=False , UpperCamelCase=0.02 , UpperCamelCase=1e-5 , UpperCamelCase=32 , **UpperCamelCase , ): super().__init__(**UpperCamelCase ) _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = patch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = embed_dim _SCREAMING_SNAKE_CASE = depths _SCREAMING_SNAKE_CASE = len(UpperCamelCase ) _SCREAMING_SNAKE_CASE = num_heads _SCREAMING_SNAKE_CASE = window_size _SCREAMING_SNAKE_CASE = mlp_ratio _SCREAMING_SNAKE_CASE = qkv_bias _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = drop_path_rate _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = use_absolute_embeddings _SCREAMING_SNAKE_CASE = layer_norm_eps _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _SCREAMING_SNAKE_CASE = int(embed_dim * 2 ** (len(UpperCamelCase ) - 1) ) _SCREAMING_SNAKE_CASE = (0, 0, 0, 0)
493
'''simple docstring''' import numpy # List of input, output pairs _snake_case : Optional[Any] = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) _snake_case : Any = (((515, 22, 13), 555), ((61, 35, 49), 150)) _snake_case : Any = [2, 4, 1, 5] _snake_case : str = len(train_data) _snake_case : int = 0.009 def _a ( _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : List[Any]="train" ): return calculate_hypothesis_value(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) - output( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _a ( _SCREAMING_SNAKE_CASE : List[Any] ): _SCREAMING_SNAKE_CASE = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def _a ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any] ): if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def _a ( _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : str ): if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def _a ( _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[int]=m ): _SCREAMING_SNAKE_CASE = 0 for i in range(_SCREAMING_SNAKE_CASE ): if index == -1: summation_value += _error(_SCREAMING_SNAKE_CASE ) else: summation_value += _error(_SCREAMING_SNAKE_CASE ) * train_data[i][0][index] return summation_value def _a ( _SCREAMING_SNAKE_CASE : Optional[Any] ): _SCREAMING_SNAKE_CASE = summation_of_cost_derivative(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) / m return cost_derivative_value def _a ( ): global parameter_vector # Tune these values to set a tolerance value for predicted output _SCREAMING_SNAKE_CASE = 0.000_002 _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 0 while True: j += 1 _SCREAMING_SNAKE_CASE = [0, 0, 0, 0] for i in range(0 , len(_SCREAMING_SNAKE_CASE ) ): _SCREAMING_SNAKE_CASE = get_cost_derivative(i - 1 ) _SCREAMING_SNAKE_CASE = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE , rtol=_SCREAMING_SNAKE_CASE , ): break _SCREAMING_SNAKE_CASE = temp_parameter_vector print(("Number of iterations:", j) ) def _a ( ): for i in range(len(_SCREAMING_SNAKE_CASE ) ): print(("Actual output value:", output(_SCREAMING_SNAKE_CASE , "test" )) ) print(("Hypothesis output:", calculate_hypothesis_value(_SCREAMING_SNAKE_CASE , "test" )) ) if __name__ == "__main__": run_gradient_descent() print("""\nTesting gradient descent for a linear hypothesis function.\n""") test_gradient_descent()
493
1
from argparse import ArgumentParser from . import BaseTransformersCLICommand def UpperCAmelCase_ ( __UpperCamelCase ): return DownloadCommand(args.model, args.cache_dir, args.force, args.trust_remote_code ) class __a ( __lowerCamelCase ): """simple docstring""" @staticmethod def __A ( _UpperCamelCase : ArgumentParser ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ =parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" ,type=_UpperCamelCase ,default=_UpperCamelCase ,help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" ,action="""store_true""" ,help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" ,action="""store_true""" ,help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" ,) download_parser.add_argument("""model""" ,type=_UpperCamelCase ,help="""Name of the model to download""" ) download_parser.set_defaults(func=_UpperCamelCase ) def __init__( self : Union[str, Any] ,_UpperCamelCase : str ,_UpperCamelCase : str ,_UpperCamelCase : bool ,_UpperCamelCase : bool ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE__ =model SCREAMING_SNAKE_CASE__ =cache SCREAMING_SNAKE_CASE__ =force SCREAMING_SNAKE_CASE__ =trust_remote_code def __A ( self : int ) -> Optional[Any]: '''simple docstring''' from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
151
def UpperCAmelCase_ ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ =len(__UpperCamelCase ), len(grid[0] ) if ( min(__UpperCamelCase, __UpperCamelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) SCREAMING_SNAKE_CASE__ =0 count += depth_first_search(__UpperCamelCase, row + 1, __UpperCamelCase, __UpperCamelCase ) count += depth_first_search(__UpperCamelCase, row - 1, __UpperCamelCase, __UpperCamelCase ) count += depth_first_search(__UpperCamelCase, __UpperCamelCase, col + 1, __UpperCamelCase ) count += depth_first_search(__UpperCamelCase, __UpperCamelCase, col - 1, __UpperCamelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
151
1
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _lowercase ( TensorFormatter[Mapping, 'torch.Tensor', Mapping] ): '''simple docstring''' def __init__( self , __UpperCamelCase=None , **__UpperCamelCase )-> List[Any]: super().__init__(features=_A ) UpperCAmelCase__ : Any = torch_tensor_kwargs import torch # noqa import torch at initialization def lowerCAmelCase__ ( self , __UpperCamelCase )-> int: import torch if isinstance(_A , _A ) and column: if all( isinstance(_A , torch.Tensor ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return torch.stack(_A ) return column def lowerCAmelCase__ ( self , __UpperCamelCase )-> int: import torch if isinstance(_A , (str, bytes, type(_A )) ): return value elif isinstance(_A , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() UpperCAmelCase__ : Any = {} if isinstance(_A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): UpperCAmelCase__ : str = {"dtype": torch.intaa} elif isinstance(_A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): UpperCAmelCase__ : Dict = {"dtype": torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(_A , PIL.Image.Image ): UpperCAmelCase__ : Optional[int] = np.asarray(_A ) return torch.tensor(_A , **{**default_dtype, **self.torch_tensor_kwargs} ) def lowerCAmelCase__ ( self , __UpperCamelCase )-> Tuple: import torch # support for torch, tf, jax etc. if hasattr(_A , "__array__" ) and not isinstance(_A , torch.Tensor ): UpperCAmelCase__ : str = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(_A , np.ndarray ): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(_A ) for substruct in data_struct] ) elif isinstance(_A , (list, tuple) ): return self._consolidate([self.recursive_tensorize(_A ) for substruct in data_struct] ) return self._tensorize(_A ) def lowerCAmelCase__ ( self , __UpperCamelCase )-> Union[str, Any]: return map_nested(self._recursive_tensorize , _A , map_list=_A ) def lowerCAmelCase__ ( self , __UpperCamelCase )-> List[Any]: UpperCAmelCase__ : List[Any] = self.numpy_arrow_extractor().extract_row(_A ) UpperCAmelCase__ : List[Any] = self.python_features_decoder.decode_row(_A ) return self.recursive_tensorize(_A ) def lowerCAmelCase__ ( self , __UpperCamelCase )-> List[Any]: UpperCAmelCase__ : List[str] = self.numpy_arrow_extractor().extract_column(_A ) UpperCAmelCase__ : Dict = self.python_features_decoder.decode_column(_A , pa_table.column_names[0] ) UpperCAmelCase__ : Any = self.recursive_tensorize(_A ) UpperCAmelCase__ : List[Any] = self._consolidate(_A ) return column def lowerCAmelCase__ ( self , __UpperCamelCase )-> Optional[Any]: UpperCAmelCase__ : Optional[int] = self.numpy_arrow_extractor().extract_batch(_A ) UpperCAmelCase__ : Optional[Any] = self.python_features_decoder.decode_batch(_A ) UpperCAmelCase__ : Dict = self.recursive_tensorize(_A ) for column_name in batch: UpperCAmelCase__ : Dict = self._consolidate(batch[column_name] ) return batch
714
"""simple docstring""" from math import isqrt def a__ ( lowerCAmelCase : int ): '''simple docstring''' UpperCAmelCase__ : Tuple = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase__ : List[Any] = False return [i for i in range(2 , lowerCAmelCase ) if is_prime[i]] def a__ ( lowerCAmelCase : int = 10**8 ): '''simple docstring''' UpperCAmelCase__ : Dict = calculate_prime_numbers(max_number // 2 ) UpperCAmelCase__ : Optional[int] = 0 UpperCAmelCase__ : Optional[int] = 0 UpperCAmelCase__ : Tuple = len(lowerCAmelCase ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(f"""{solution() = }""")
660
0
from math import factorial, pi def lowerCamelCase__ ( snake_case_ : float , snake_case_ : int = 30 ) -> float: if not isinstance(SCREAMING_SNAKE_CASE_ , (int, float) ): raise ValueError('''maclaurin_sin() requires either an int or float for theta''' ) if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) or accuracy <= 0: raise ValueError('''maclaurin_sin() requires a positive int for accuracy''' ) __snake_case = float(SCREAMING_SNAKE_CASE_ ) __snake_case = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(SCREAMING_SNAKE_CASE_ ) ) def lowerCamelCase__ ( snake_case_ : float , snake_case_ : int = 30 ) -> float: if not isinstance(SCREAMING_SNAKE_CASE_ , (int, float) ): raise ValueError('''maclaurin_cos() requires either an int or float for theta''' ) if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) or accuracy <= 0: raise ValueError('''maclaurin_cos() requires a positive int for accuracy''' ) __snake_case = float(SCREAMING_SNAKE_CASE_ ) __snake_case = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(10)) print(maclaurin_sin(-10)) print(maclaurin_sin(10, 15)) print(maclaurin_sin(-10, 15)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(10, 15)) print(maclaurin_cos(-10, 15))
592
from queue import PriorityQueue from typing import Any import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: dict , SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: set , SCREAMING_SNAKE_CASE_: set , SCREAMING_SNAKE_CASE_: dict , SCREAMING_SNAKE_CASE_: dict , SCREAMING_SNAKE_CASE_: PriorityQueue , SCREAMING_SNAKE_CASE_: dict , SCREAMING_SNAKE_CASE_: float | int , ) -> float | int: '''simple docstring''' for nxt, d in graph[v]: if nxt in visited_forward: continue A__ = cst_fwd.get(SCREAMING_SNAKE_CASE_ , np.inf ) A__ = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) A__ = new_cost_f A__ = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: A__ = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: dict , SCREAMING_SNAKE_CASE_: dict ) -> int: '''simple docstring''' A__ = -1 A__ = set() A__ = set() A__ = {source: 0} A__ = {destination: 0} A__ = {source: None} A__ = {destination: None} A__ = PriorityQueue() A__ = PriorityQueue() A__ = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): A__ , A__ = queue_forward.get() visited_forward.add(SCREAMING_SNAKE_CASE_ ) A__ , A__ = queue_backward.get() visited_backward.add(SCREAMING_SNAKE_CASE_ ) A__ = pass_and_relaxation( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) A__ = pass_and_relaxation( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: A__ = shortest_distance return shortest_path_distance lowerCAmelCase__ = { """B""": [["""C""", 1]], """C""": [["""D""", 1]], """D""": [["""F""", 1]], """E""": [["""B""", 1], ["""G""", 2]], """F""": [], """G""": [["""F""", 1]], } lowerCAmelCase__ = { """B""": [["""E""", 1]], """C""": [["""B""", 1]], """D""": [["""C""", 1]], """F""": [["""D""", 1], ["""G""", 1]], """E""": [[None, np.inf]], """G""": [["""E""", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
514
0
import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) def _UpperCAmelCase ( A , A ): '''simple docstring''' UpperCAmelCase__ =nn.functional.normalize(A ) UpperCAmelCase__ =nn.functional.normalize(A ) return torch.mm(A , normalized_text_embeds.t() ) class snake_case_ ( a ): __UpperCamelCase = CLIPConfig __UpperCamelCase = ['CLIPEncoderLayer'] def __init__( self, A_ ) -> Tuple: super().__init__(A_ ) UpperCAmelCase__ =CLIPVisionModel(config.vision_config ) UpperCAmelCase__ =nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=A_ ) UpperCAmelCase__ =nn.Parameter(torch.ones(17, config.projection_dim ), requires_grad=A_ ) UpperCAmelCase__ =nn.Parameter(torch.ones(3, config.projection_dim ), requires_grad=A_ ) UpperCAmelCase__ =nn.Parameter(torch.ones(17 ), requires_grad=A_ ) UpperCAmelCase__ =nn.Parameter(torch.ones(3 ), requires_grad=A_ ) @torch.no_grad() def __UpperCAmelCase ( self, A_, A_ ) -> List[str]: UpperCAmelCase__ =self.vision_model(A_ )[1] # pooled_output UpperCAmelCase__ =self.visual_projection(A_ ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 UpperCAmelCase__ =cosine_distance(A_, self.special_care_embeds ).cpu().float().numpy() UpperCAmelCase__ =cosine_distance(A_, self.concept_embeds ).cpu().float().numpy() UpperCAmelCase__ =[] UpperCAmelCase__ =image_embeds.shape[0] for i in range(A_ ): UpperCAmelCase__ ={"special_scores": {}, "special_care": [], "concept_scores": {}, "bad_concepts": []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images UpperCAmelCase__ =0.0 for concept_idx in range(len(special_cos_dist[0] ) ): UpperCAmelCase__ =special_cos_dist[i][concept_idx] UpperCAmelCase__ =self.special_care_embeds_weights[concept_idx].item() UpperCAmelCase__ =round(concept_cos - concept_threshold + adjustment, 3 ) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img["special_scores"][concept_idx]} ) UpperCAmelCase__ =0.01 for concept_idx in range(len(cos_dist[0] ) ): UpperCAmelCase__ =cos_dist[i][concept_idx] UpperCAmelCase__ =self.concept_embeds_weights[concept_idx].item() UpperCAmelCase__ =round(concept_cos - concept_threshold + adjustment, 3 ) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(A_ ) result.append(A_ ) UpperCAmelCase__ =[len(res["bad_concepts"] ) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def __UpperCAmelCase ( self, A_, A_ ) -> List[Any]: UpperCAmelCase__ =self.vision_model(A_ )[1] # pooled_output UpperCAmelCase__ =self.visual_projection(A_ ) UpperCAmelCase__ =cosine_distance(A_, self.special_care_embeds ) UpperCAmelCase__ =cosine_distance(A_, self.concept_embeds ) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images UpperCAmelCase__ =0.0 UpperCAmelCase__ =special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) UpperCAmelCase__ =torch.any(special_scores > 0, dim=1 ) UpperCAmelCase__ =special_care * 0.01 UpperCAmelCase__ =special_adjustment.unsqueeze(1 ).expand(-1, cos_dist.shape[1] ) UpperCAmelCase__ =(cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) UpperCAmelCase__ =torch.any(concept_scores > 0, dim=1 ) return images, has_nsfw_concepts
703
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case_ ( a ): '''simple docstring''' __UpperCamelCase = ['image_processor', 'tokenizer'] __UpperCamelCase = 'BridgeTowerImageProcessor' __UpperCamelCase = ('RobertaTokenizer', 'RobertaTokenizerFast') def __init__( self, A_, A_ ) -> Dict: super().__init__(A_, A_ ) def __call__( self, A_, A_ = None, A_ = True, A_ = False, A_ = None, A_ = None, A_ = 0, A_ = None, A_ = None, A_ = None, A_ = False, A_ = False, A_ = False, A_ = False, A_ = True, A_ = None, **A_, ) -> BatchEncoding: UpperCAmelCase__ =self.tokenizer( text=A_, add_special_tokens=A_, padding=A_, truncation=A_, max_length=A_, stride=A_, pad_to_multiple_of=A_, return_token_type_ids=A_, return_attention_mask=A_, return_overflowing_tokens=A_, return_special_tokens_mask=A_, return_offsets_mapping=A_, return_length=A_, verbose=A_, return_tensors=A_, **A_, ) # add pixel_values + pixel_mask UpperCAmelCase__ =self.image_processor( A_, return_tensors=A_, do_normalize=A_, do_center_crop=A_, **A_ ) encoding.update(A_ ) return encoding def __UpperCAmelCase ( self, *A_, **A_ ) -> Any: return self.tokenizer.batch_decode(*A_, **A_ ) def __UpperCAmelCase ( self, *A_, **A_ ) -> Tuple: return self.tokenizer.decode(*A_, **A_ ) @property def __UpperCAmelCase ( self ) -> Union[str, Any]: UpperCAmelCase__ =self.tokenizer.model_input_names UpperCAmelCase__ =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
510
0
def UpperCAmelCase_ ( __UpperCAmelCase : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: SCREAMING_SNAKE_CASE_ = set() # Replace all the whitespace in our sentence SCREAMING_SNAKE_CASE_ = input_str.replace(' ' , '' ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(__UpperCAmelCase ) == 26 def UpperCAmelCase_ ( __UpperCAmelCase : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: SCREAMING_SNAKE_CASE_ = [False] * 26 for char in input_str: if char.islower(): SCREAMING_SNAKE_CASE_ = True elif char.isupper(): SCREAMING_SNAKE_CASE_ = True return all(__UpperCAmelCase ) def UpperCAmelCase_ ( __UpperCAmelCase : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def UpperCAmelCase_ ( ) -> None: from timeit import timeit SCREAMING_SNAKE_CASE_ = 'from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest' print(timeit('is_pangram()' , setup=__UpperCAmelCase ) ) print(timeit('is_pangram_faster()' , setup=__UpperCAmelCase ) ) print(timeit('is_pangram_fastest()' , setup=__UpperCAmelCase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
31
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def UpperCAmelCase ( A : dict , A : str , A : set , A : set , A : dict , A : dict , A : PriorityQueue , A : dict , A : float | int , ): for nxt, d in graph[v]: if nxt in visited_forward: continue SCREAMING_SNAKE_CASE : str = cst_fwd.get(A , np.inf ) SCREAMING_SNAKE_CASE : Optional[int] = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) SCREAMING_SNAKE_CASE : Tuple = new_cost_f SCREAMING_SNAKE_CASE : List[str] = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: SCREAMING_SNAKE_CASE : Union[str, Any] = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def UpperCAmelCase ( A : str , A : str , A : dict , A : dict ): SCREAMING_SNAKE_CASE : Dict = -1 SCREAMING_SNAKE_CASE : Optional[int] = set() SCREAMING_SNAKE_CASE : Optional[Any] = set() SCREAMING_SNAKE_CASE : Any = {source: 0} SCREAMING_SNAKE_CASE : Tuple = {destination: 0} SCREAMING_SNAKE_CASE : Union[str, Any] = {source: None} SCREAMING_SNAKE_CASE : Optional[int] = {destination: None} SCREAMING_SNAKE_CASE : PriorityQueue[Any] = PriorityQueue() SCREAMING_SNAKE_CASE : PriorityQueue[Any] = PriorityQueue() SCREAMING_SNAKE_CASE : List[str] = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : str = queue_forward.get() visited_forward.add(A ) SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : int = queue_backward.get() visited_backward.add(A ) SCREAMING_SNAKE_CASE : Tuple = pass_and_relaxation( A , A , A , A , A , A , A , A , A , ) SCREAMING_SNAKE_CASE : Optional[int] = pass_and_relaxation( A , A , A , A , A , A , A , A , A , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: SCREAMING_SNAKE_CASE : int = shortest_distance return shortest_path_distance lowerCAmelCase_ : int = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } lowerCAmelCase_ : str = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
527
0
'''simple docstring''' from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCAmelCase__ ( UpperCamelCase__ ): """simple docstring""" def _UpperCAmelCase ( self: List[str] ) -> int: '''simple docstring''' __UpperCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_a , "embed_dim" ) ) self.parent.assertTrue(hasattr(_a , "num_heads" ) ) class UpperCAmelCase__ : """simple docstring""" def __init__( self: Union[str, Any] , __lowerCAmelCase: List[str] , __lowerCAmelCase: List[str]=13 , __lowerCAmelCase: Any=64 , __lowerCAmelCase: List[Any]=3 , __lowerCAmelCase: List[str]=[16, 48, 96] , __lowerCAmelCase: Optional[Any]=[1, 3, 6] , __lowerCAmelCase: List[Any]=[1, 2, 10] , __lowerCAmelCase: Any=[7, 3, 3] , __lowerCAmelCase: List[Any]=[4, 2, 2] , __lowerCAmelCase: List[Any]=[2, 1, 1] , __lowerCAmelCase: Dict=[2, 2, 2] , __lowerCAmelCase: Union[str, Any]=[False, False, True] , __lowerCAmelCase: Optional[Any]=[0.0, 0.0, 0.0] , __lowerCAmelCase: Dict=0.02 , __lowerCAmelCase: Tuple=1E-12 , __lowerCAmelCase: Optional[Any]=True , __lowerCAmelCase: Optional[Any]=True , __lowerCAmelCase: Union[str, Any]=2 , ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase = parent __UpperCAmelCase = batch_size __UpperCAmelCase = image_size __UpperCAmelCase = patch_sizes __UpperCAmelCase = patch_stride __UpperCAmelCase = patch_padding __UpperCAmelCase = is_training __UpperCAmelCase = use_labels __UpperCAmelCase = num_labels __UpperCAmelCase = num_channels __UpperCAmelCase = embed_dim __UpperCAmelCase = num_heads __UpperCAmelCase = stride_kv __UpperCAmelCase = depth __UpperCAmelCase = cls_token __UpperCAmelCase = attention_drop_rate __UpperCAmelCase = initializer_range __UpperCAmelCase = layer_norm_eps def _UpperCAmelCase ( self: str ) -> Dict: '''simple docstring''' __UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase = None if self.use_labels: # create a random int32 tensor of given shape __UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) __UpperCAmelCase = self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self: Dict ) -> List[Any]: '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self: Tuple , __lowerCAmelCase: Tuple , __lowerCAmelCase: Any , __lowerCAmelCase: Dict ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase = TFCvtModel(config=_a ) __UpperCAmelCase = model(_a , training=_a ) __UpperCAmelCase = (self.image_size, self.image_size) __UpperCAmelCase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __UpperCAmelCase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __UpperCAmelCase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def _UpperCAmelCase ( self: Union[str, Any] , __lowerCAmelCase: List[str] , __lowerCAmelCase: List[Any] , __lowerCAmelCase: Optional[int] ) -> List[str]: '''simple docstring''' __UpperCAmelCase = self.num_labels __UpperCAmelCase = TFCvtForImageClassification(_a ) __UpperCAmelCase = model(_a , labels=_a , training=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self: List[str] ) -> Tuple: '''simple docstring''' __UpperCAmelCase = self.prepare_config_and_inputs() __UpperCAmelCase = config_and_inputs __UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ : int = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () lowerCAmelCase__ : int = ( {'feature-extraction': TFCvtModel, 'image-classification': TFCvtForImageClassification} if is_tf_available() else {} ) lowerCAmelCase__ : Dict = False lowerCAmelCase__ : Optional[int] = False lowerCAmelCase__ : int = False lowerCAmelCase__ : str = False lowerCAmelCase__ : Tuple = False def _UpperCAmelCase ( self: List[Any] ) -> List[Any]: '''simple docstring''' __UpperCAmelCase = TFCvtModelTester(self ) __UpperCAmelCase = TFCvtConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def _UpperCAmelCase ( self: Dict ) -> str: '''simple docstring''' self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="Cvt does not output attentions" ) def _UpperCAmelCase ( self: Union[str, Any] ) -> List[str]: '''simple docstring''' pass @unittest.skip(reason="Cvt does not use inputs_embeds" ) def _UpperCAmelCase ( self: List[str] ) -> List[Any]: '''simple docstring''' pass @unittest.skip(reason="Cvt does not support input and output embeddings" ) def _UpperCAmelCase ( self: List[Any] ) -> str: '''simple docstring''' pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("GPU" ) ) == 0 , reason="TF does not support backprop for grouped convolutions on CPU." , ) def _UpperCAmelCase ( self: Dict ) -> Union[str, Any]: '''simple docstring''' super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("GPU" ) ) == 0 , reason="TF does not support backprop for grouped convolutions on CPU." , ) @slow def _UpperCAmelCase ( self: Tuple ) -> List[str]: '''simple docstring''' super().test_keras_fit() @unittest.skip(reason="Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8" ) def _UpperCAmelCase ( self: Any ) -> int: '''simple docstring''' __UpperCAmelCase = tf.keras.mixed_precision.Policy("mixed_float16" ) tf.keras.mixed_precision.set_global_policy(_a ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("float32" ) def _UpperCAmelCase ( self: str ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase = model_class(_a ) __UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase = [*signature.parameters.keys()] __UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _a ) def _UpperCAmelCase ( self: int ) -> Optional[Any]: '''simple docstring''' def check_hidden_states_output(__lowerCAmelCase: Any , __lowerCAmelCase: Union[str, Any] , __lowerCAmelCase: Union[str, Any] ): __UpperCAmelCase = model_class(_a ) __UpperCAmelCase = model(**self._prepare_for_class(_a , _a ) ) __UpperCAmelCase = outputs.hidden_states __UpperCAmelCase = len(self.model_tester.depth ) self.assertEqual(len(_a ) , _a ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase = True check_hidden_states_output(_a , _a , _a ) def _UpperCAmelCase ( self: Optional[Any] ) -> Dict: '''simple docstring''' __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def _UpperCAmelCase ( self: Optional[Any] ) -> Tuple: '''simple docstring''' __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def _UpperCAmelCase ( self: Optional[int] ) -> Any: '''simple docstring''' for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase = TFCvtModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowerCAmelCase ( ) -> Dict: __UpperCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def _UpperCAmelCase ( self: Union[str, Any] ) -> int: '''simple docstring''' return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def _UpperCAmelCase ( self: Any ) -> Tuple: '''simple docstring''' __UpperCAmelCase = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __UpperCAmelCase = self.default_image_processor __UpperCAmelCase = prepare_img() __UpperCAmelCase = image_processor(images=_a , return_tensors="tf" ) # forward pass __UpperCAmelCase = model(**_a ) # verify the logits __UpperCAmelCase = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , _a ) __UpperCAmelCase = tf.constant([0.9285, 0.9015, -0.3150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _a , atol=1E-4 ) )
710
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCAmelCase__ ( snake_case , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ : int = DiTPipeline lowerCAmelCase__ : Any = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS lowerCAmelCase__ : Optional[Any] = PipelineTesterMixin.required_optional_params - { 'latents', 'num_images_per_prompt', 'callback', 'callback_steps', } lowerCAmelCase__ : List[Any] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS lowerCAmelCase__ : Optional[int] = False def _UpperCAmelCase ( self: str ) -> List[Any]: '''simple docstring''' torch.manual_seed(0 ) __UpperCAmelCase = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=__lowerCAmelCase , activation_fn="gelu-approximate" , num_embeds_ada_norm=1_000 , norm_type="ada_norm_zero" , norm_elementwise_affine=__lowerCAmelCase , ) __UpperCAmelCase = AutoencoderKL() __UpperCAmelCase = DDIMScheduler() __UpperCAmelCase = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def _UpperCAmelCase ( self: str , __lowerCAmelCase: str , __lowerCAmelCase: Optional[int]=0 ) -> int: '''simple docstring''' if str(__lowerCAmelCase ).startswith("mps" ): __UpperCAmelCase = torch.manual_seed(__lowerCAmelCase ) else: __UpperCAmelCase = torch.Generator(device=__lowerCAmelCase ).manual_seed(__lowerCAmelCase ) __UpperCAmelCase = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def _UpperCAmelCase ( self: Tuple ) -> Dict: '''simple docstring''' __UpperCAmelCase = "cpu" __UpperCAmelCase = self.get_dummy_components() __UpperCAmelCase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) __UpperCAmelCase = self.get_dummy_inputs(__lowerCAmelCase ) __UpperCAmelCase = pipe(**__lowerCAmelCase ).images __UpperCAmelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) __UpperCAmelCase = np.array([0.2946, 0.6601, 0.4329, 0.3296, 0.4144, 0.5319, 0.7273, 0.5013, 0.4457] ) __UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__lowerCAmelCase , 1E-3 ) def _UpperCAmelCase ( self: int ) -> str: '''simple docstring''' self._test_inference_batch_single_identical(relax_max_difference=__lowerCAmelCase , expected_max_diff=1E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _UpperCAmelCase ( self: Optional[int] ) -> str: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @require_torch_gpu @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def _UpperCAmelCase ( self: Any ) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self: Dict ) -> List[Any]: '''simple docstring''' __UpperCAmelCase = torch.manual_seed(0 ) __UpperCAmelCase = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) __UpperCAmelCase = ["vase", "umbrella", "white shark", "white wolf"] __UpperCAmelCase = pipe.get_label_ids(__lowerCAmelCase ) __UpperCAmelCase = pipe(__lowerCAmelCase , generator=__lowerCAmelCase , num_inference_steps=40 , output_type="np" ).images for word, image in zip(__lowerCAmelCase , __lowerCAmelCase ): __UpperCAmelCase = load_numpy( F'''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy''' ) assert np.abs((expected_image - image).max() ) < 1E-2 def _UpperCAmelCase ( self: Any ) -> List[str]: '''simple docstring''' __UpperCAmelCase = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) __UpperCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) __UpperCAmelCase = ["vase", "umbrella"] __UpperCAmelCase = pipe.get_label_ids(__lowerCAmelCase ) __UpperCAmelCase = torch.manual_seed(0 ) __UpperCAmelCase = pipe(__lowerCAmelCase , generator=__lowerCAmelCase , num_inference_steps=25 , output_type="np" ).images for word, image in zip(__lowerCAmelCase , __lowerCAmelCase ): __UpperCAmelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" F'''/dit/{word}_512.npy''' ) assert np.abs((expected_image - image).max() ) < 1E-1
286
0
def lowerCamelCase__ ( _a): if a < 0: raise ValueError("Input value must be a positive integer") elif isinstance(_a , _a): raise TypeError("Input value must be a 'int' type") return bin(_a).count("1") if __name__ == "__main__": import doctest doctest.testmod()
25
'''simple docstring''' import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCAmelCase_ ( snake_case__ ): """simple docstring""" def __init__( self : List[Any] , SCREAMING_SNAKE_CASE__ : NestedDataStructureLike[PathLike] , SCREAMING_SNAKE_CASE__ : Optional[NamedSplit] = None , SCREAMING_SNAKE_CASE__ : Optional[Features] = None , SCREAMING_SNAKE_CASE__ : str = None , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Optional[str] = None , SCREAMING_SNAKE_CASE__ : Optional[int] = None , **SCREAMING_SNAKE_CASE__ : Tuple , ): '''simple docstring''' super().__init__( SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ , streaming=SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) __a = field __a = path_or_paths if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else {self.split: path_or_paths} __a = Json( cache_dir=SCREAMING_SNAKE_CASE__ , data_files=SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , field=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) def __a ( self : Dict ): '''simple docstring''' if self.streaming: __a = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __a = None __a = None __a = None __a = None self.builder.download_and_prepare( download_config=SCREAMING_SNAKE_CASE__ , download_mode=SCREAMING_SNAKE_CASE__ , verification_mode=SCREAMING_SNAKE_CASE__ , base_path=SCREAMING_SNAKE_CASE__ , num_proc=self.num_proc , ) __a = self.builder.as_dataset( split=self.split , verification_mode=SCREAMING_SNAKE_CASE__ , in_memory=self.keep_in_memory ) return dataset class lowerCAmelCase_ : """simple docstring""" def __init__( self : str , SCREAMING_SNAKE_CASE__ : Dataset , SCREAMING_SNAKE_CASE__ : Union[PathLike, BinaryIO] , SCREAMING_SNAKE_CASE__ : Optional[int] = None , SCREAMING_SNAKE_CASE__ : Optional[int] = None , **SCREAMING_SNAKE_CASE__ : List[str] , ): '''simple docstring''' if num_proc is not None and num_proc <= 0: raise ValueError(f'''num_proc {num_proc} must be an integer > 0.''' ) __a = dataset __a = path_or_buf __a = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __a = num_proc __a = """utf-8""" __a = to_json_kwargs def __a ( self : int ): '''simple docstring''' __a = self.to_json_kwargs.pop("""path_or_buf""" , SCREAMING_SNAKE_CASE__ ) __a = self.to_json_kwargs.pop("""orient""" , """records""" ) __a = self.to_json_kwargs.pop("""lines""" , True if orient == """records""" else False ) __a = self.to_json_kwargs.pop("""index""" , False if orient in ["""split""", """table"""] else True ) __a = self.to_json_kwargs.pop("""compression""" , SCREAMING_SNAKE_CASE__ ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(f'''`datasets` currently does not support {compression} compression''' ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , """wb""" , compression=SCREAMING_SNAKE_CASE__ ) as buffer: __a = self._write(file_obj=SCREAMING_SNAKE_CASE__ , orient=SCREAMING_SNAKE_CASE__ , lines=SCREAMING_SNAKE_CASE__ , index=SCREAMING_SNAKE_CASE__ , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( f'''The compression parameter is not supported when writing to a buffer, but compression={compression}''' """ was passed. Please provide a local path instead.""" ) __a = self._write( file_obj=self.path_or_buf , orient=SCREAMING_SNAKE_CASE__ , lines=SCREAMING_SNAKE_CASE__ , index=SCREAMING_SNAKE_CASE__ , **self.to_json_kwargs ) return written def __a ( self : List[Any] , SCREAMING_SNAKE_CASE__ : List[Any] ): '''simple docstring''' __a , __a , __a , __a , __a = args __a = query_table( table=self.dataset.data , key=slice(SCREAMING_SNAKE_CASE__ , offset + self.batch_size ) , indices=self.dataset._indices , ) __a = batch.to_pandas().to_json( path_or_buf=SCREAMING_SNAKE_CASE__ , orient=SCREAMING_SNAKE_CASE__ , lines=SCREAMING_SNAKE_CASE__ , index=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if not json_str.endswith("""\n""" ): json_str += "\n" return json_str.encode(self.encoding ) def __a ( self : Tuple , SCREAMING_SNAKE_CASE__ : BinaryIO , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[str] , **SCREAMING_SNAKE_CASE__ : Optional[int] , ): '''simple docstring''' __a = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating json from Arrow format""" , ): __a = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(SCREAMING_SNAKE_CASE__ ) else: __a , __a = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating json from Arrow format""" , ): written += file_obj.write(SCREAMING_SNAKE_CASE__ ) return written
582
0
"""simple docstring""" def lowercase_ ( _UpperCAmelCase ): """simple docstring""" return credit_card_number.startswith(('''34''', '''35''', '''37''', '''4''', '''5''', '''6''') ) def lowercase_ ( _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = credit_card_number A_ : Dict = 0 A_ : List[str] = len(_UpperCAmelCase ) - 2 for i in range(_UpperCAmelCase , -1 , -2 ): # double the value of every second digit A_ : Optional[Any] = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 A_ : Tuple = cc_number[:i] + str(_UpperCAmelCase ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(_UpperCAmelCase ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def lowercase_ ( _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = f"""{credit_card_number} is an invalid credit card number because""" if not credit_card_number.isdigit(): print(f"""{error_message} it has nonnumerical characters.""" ) return False if not 13 <= len(_UpperCAmelCase ) <= 16: print(f"""{error_message} of its length.""" ) return False if not validate_initial_digits(_UpperCAmelCase ): print(f"""{error_message} of its first two digits.""" ) return False if not luhn_validation(_UpperCAmelCase ): print(f"""{error_message} it fails the Luhn check.""" ) return False print(f"""{credit_card_number} is a valid credit card number.""" ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
361
"""simple docstring""" import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _lowerCamelCase : int = logging.getLogger(__name__) _lowerCamelCase : List[Any] = 'pytorch_model.bin' @dataclasses.dataclass class lowercase : __lowerCAmelCase : str = dataclasses.field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models."""}) __lowerCAmelCase : Optional[str] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co."""} , ) @dataclasses.dataclass class lowercase : __lowerCAmelCase : str = dataclasses.field(metadata={"""help""": """A csv or a json file containing the training data."""}) __lowerCAmelCase : str = dataclasses.field(metadata={"""help""": """A csv or a json file containing the data to predict on."""}) __lowerCAmelCase : Optional[str] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """A csv or a json file containing the validation data."""}) __lowerCAmelCase : Optional[str] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """The name of the task to train on."""} , ) __lowerCAmelCase : Optional[List[str]] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """The list of labels for the task."""}) @dataclasses.dataclass class lowercase : __lowerCAmelCase : str = dataclasses.field( metadata={"""help""": """The output directory where the model predictions and checkpoints will be written."""}) __lowerCAmelCase : Optional[str] = dataclasses.field( default="""accuracy""" , metadata={"""help""": """The evaluation metric used for the task."""}) __lowerCAmelCase : Optional[str] = dataclasses.field( default="""no""" , metadata={ """help""": """The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]""" } , ) __lowerCAmelCase : Optional[int] = dataclasses.field( default=10 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) __lowerCAmelCase : Optional[float] = dataclasses.field( default=0.0 , metadata={ """help""": """How much the specified evaluation metric must improve to satisfy early stopping conditions.""" } , ) __lowerCAmelCase : Optional[bool] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the confidence score."""} , ) __lowerCAmelCase : Optional[bool] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the validation performance."""} , ) __lowerCAmelCase : Optional[bool] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """Whether to fine-tune on labeled data after pseudo training."""} , ) __lowerCAmelCase : Optional[float] = dataclasses.field( default=0.0 , metadata={"""help""": """Confidence threshold for pseudo-labeled data filtering."""} , ) __lowerCAmelCase : Optional[int] = dataclasses.field( default=100 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) __lowerCAmelCase : Optional[int] = dataclasses.field( default=__UpperCAmelCase , metadata={"""help""": """Random seed for initialization."""} , ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: A_ : Optional[int] = dataset.filter(lambda _UpperCAmelCase : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 A_ : Optional[Any] = int(eval_result * len(_UpperCAmelCase ) ) print(_UpperCAmelCase ) A_ : str = dataset.sort('''probability''' , reverse=_UpperCAmelCase ) A_ : str = dataset.select(range(_UpperCAmelCase ) ) A_ : int = dataset.remove_columns(['''label''', '''probability'''] ) A_ : int = dataset.rename_column('''prediction''' , '''label''' ) A_ : int = dataset.map(lambda _UpperCAmelCase : {"label": idalabel[example["label"]]} ) A_ : List[Any] = dataset.shuffle(seed=args.seed ) A_ : str = os.path.join(_UpperCAmelCase , f"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(_UpperCAmelCase , index=_UpperCAmelCase ) else: dataset.to_json(_UpperCAmelCase ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" A_ : Any = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() A_ : str = STModelArguments(model_name_or_path=_UpperCAmelCase ) A_ : List[str] = STDataArguments(train_file=_UpperCAmelCase , infer_file=_UpperCAmelCase ) A_ : Any = STTrainingArguments(output_dir=_UpperCAmelCase ) A_ : int = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(_UpperCAmelCase ).items(): setattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) for key, value in kwargs.items(): if hasattr(_UpperCAmelCase , _UpperCAmelCase ): setattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Sanity checks A_ : Any = {} A_ : Dict = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None A_ : str = args.train_file A_ : str = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None A_ : List[str] = args.eval_file for key in data_files: A_ : Dict = data_files[key].split('''.''' )[-1] assert extension in ["csv", "json"], f"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: A_ : Optional[Any] = extension else: assert extension == args.data_file_extension, f"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), f"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info('''Creating the initial data directory for self-training...''' ) A_ : List[Any] = f"""{args.output_dir}/self-train_iter-{{}}""".format A_ : Union[str, Any] = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=_UpperCAmelCase ) os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) accelerator.wait_for_everyone() A_ : Optional[int] = None A_ : Union[str, Any] = None A_ : Optional[int] = 0 A_ : List[Any] = False # Show the progress bar A_ : Union[str, Any] = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): A_ : Dict = data_dir_format(_UpperCAmelCase ) assert os.path.exists(_UpperCAmelCase ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 A_ : int = os.path.join(_UpperCAmelCase , '''stage-1''' ) A_ : List[Any] = { '''accelerator''': accelerator, '''model_name_or_path''': args.model_name_or_path, '''cache_dir''': args.cache_dir, '''do_train''': True, '''train_file''': data_files['''train'''] if iteration == 0 else data_files['''train_pseudo'''], '''do_eval''': True if args.eval_file is not None else False, '''eval_file''': data_files['''eval'''], '''do_predict''': True, '''infer_file''': data_files['''infer'''], '''task_name''': args.task_name, '''label_list''': args.label_list, '''output_dir''': current_output_dir, '''eval_metric''': args.eval_metric, '''evaluation_strategy''': args.evaluation_strategy, '''early_stopping_patience''': args.early_stopping_patience, '''early_stopping_threshold''': args.early_stopping_threshold, '''seed''': args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(_UpperCAmelCase , _UpperCAmelCase ): arguments_dict.update({key: value} ) A_ : Any = os.path.join(_UpperCAmelCase , '''best-checkpoint''' , _UpperCAmelCase ) if os.path.exists(_UpperCAmelCase ): logger.info( '''Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.''' , _UpperCAmelCase , _UpperCAmelCase , ) else: logger.info('''***** Running self-training: iteration: %d, stage: 1 *****''' , _UpperCAmelCase ) finetune(**_UpperCAmelCase ) accelerator.wait_for_everyone() assert os.path.exists(_UpperCAmelCase ) logger.info('''Self-training job completed: iteration: %d, stage: 1.''' , _UpperCAmelCase ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data A_ : List[Any] = os.path.join(_UpperCAmelCase , '''best-checkpoint''' ) A_ : Optional[int] = os.path.join(_UpperCAmelCase , '''stage-2''' ) # Update arguments_dict A_ : int = model_path A_ : Optional[Any] = data_files['''train'''] A_ : Optional[int] = current_output_dir A_ : str = os.path.join(_UpperCAmelCase , '''best-checkpoint''' , _UpperCAmelCase ) if os.path.exists(_UpperCAmelCase ): logger.info( '''Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.''' , _UpperCAmelCase , _UpperCAmelCase , ) else: logger.info('''***** Running self-training: iteration: %d, stage: 2 *****''' , _UpperCAmelCase ) finetune(**_UpperCAmelCase ) accelerator.wait_for_everyone() assert os.path.exists(_UpperCAmelCase ) logger.info('''Self-training job completed: iteration: %d, stage: 2.''' , _UpperCAmelCase ) A_ : Dict = iteration A_ : List[Any] = data_dir_format(iteration + 1 ) A_ : str = AutoConfig.from_pretrained(os.path.join(_UpperCAmelCase , '''best-checkpoint''' ) ) A_ : List[Any] = config.idalabel A_ : Any = os.path.join(_UpperCAmelCase , '''eval_results_best-checkpoint.json''' ) A_ : Union[str, Any] = os.path.join(_UpperCAmelCase , '''test_results_best-checkpoint.json''' ) assert os.path.exists(_UpperCAmelCase ) with open(_UpperCAmelCase , '''r''' ) as f: A_ : Union[str, Any] = float(json.load(_UpperCAmelCase )[args.eval_metric] ) A_ : Union[str, Any] = os.path.join(_UpperCAmelCase , '''infer_output_best-checkpoint.csv''' ) assert os.path.exists(_UpperCAmelCase ) # Loading the dataset from local csv or json files. A_ : List[str] = load_dataset(args.data_file_extension , data_files={'''data''': data_files['''infer''']} )['''data'''] A_ : Any = load_dataset('''csv''' , data_files={'''data''': infer_output_file} )['''data'''] if accelerator.is_main_process: os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) shutil.copy(_UpperCAmelCase , os.path.join(_UpperCAmelCase , f"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(_UpperCAmelCase ): shutil.copy(_UpperCAmelCase , os.path.join(_UpperCAmelCase , f"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) accelerator.wait_for_everyone() A_ : int = os.path.join(_UpperCAmelCase , f"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: A_ : Union[str, Any] = eval_result if best_iteration is None: A_ : str = new_iteration A_ : Dict = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: A_ : int = new_iteration A_ : Tuple = new_eval_result A_ : Union[str, Any] = 0 else: if new_eval_result == best_eval_result: A_ : Union[str, Any] = new_iteration A_ : List[str] = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: A_ : int = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info('''Best iteration: %d''' , _UpperCAmelCase ) logger.info('''Best evaluation result: %s = %f''' , args.eval_metric , _UpperCAmelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_UpperCAmelCase , f"""eval_results_iter-{iteration}.json""" ) , os.path.join(_UpperCAmelCase , '''eval_results_best-iteration.json''' ) , ) else: # Assume that the last iteration is the best logger.info('''Best iteration: %d''' , args.max_selftrain_iterations - 1 ) logger.info('''Best evaluation result: %s = %f''' , args.eval_metric , _UpperCAmelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_UpperCAmelCase , f"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(_UpperCAmelCase , '''eval_results_best-iteration.json''' ) , )
361
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , __A : Any , __A : List[str]=7 , __A : int=3 , __A : Tuple=1_8 , __A : Any=3_0 , __A : Tuple=4_0_0 , __A : Optional[Any]=True , __A : List[Any]=None , __A : List[str]=True , __A : str=None , __A : Union[str, Any]=True , __A : int=[0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] , __A : Any=[0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] , __A : Optional[Any]=True , ): snake_case__ : Dict = size if size is not None else {"height": 2_2_4, "width": 2_2_4} snake_case__ : int = crop_size if crop_size is not None else {"height": 1_8, "width": 1_8} snake_case__ : List[str] = parent snake_case__ : int = batch_size snake_case__ : Optional[Any] = num_channels snake_case__ : Tuple = image_size snake_case__ : Optional[Any] = min_resolution snake_case__ : Union[str, Any] = max_resolution snake_case__ : Tuple = do_resize snake_case__ : Dict = size snake_case__ : str = do_center_crop snake_case__ : Optional[int] = crop_size snake_case__ : str = do_normalize snake_case__ : Optional[Any] = image_mean snake_case__ : int = image_std snake_case__ : List[str] = do_convert_rgb def _lowercase ( self : Tuple ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def _lowercase ( self : List[Any] , __A : Optional[int]=False , __A : List[str]=False , __A : int=False ): assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: snake_case__ : List[Any] = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 2_5_5 , size=(self.num_channels, self.max_resolution, self.max_resolution) , dtype=np.uinta ) ) else: snake_case__ : str = [] for i in range(self.batch_size ): snake_case__, snake_case__ : Union[str, Any] = np.random.choice(np.arange(self.min_resolution , self.max_resolution ) , 2 ) image_inputs.append(np.random.randint(2_5_5 , size=(self.num_channels, width, height) , dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension snake_case__ : Dict = [Image.fromarray(np.moveaxis(__A , 0 , -1 ) ) for x in image_inputs] if torchify: snake_case__ : Optional[int] = [torch.from_numpy(__A ) for x in image_inputs] return image_inputs @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( __UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = ChineseCLIPImageProcessor if is_vision_available() else None def _lowercase ( self : int ): snake_case__ : Optional[Any] = ChineseCLIPImageProcessingTester(self , do_center_crop=__A ) @property def _lowercase ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : Union[str, Any] ): snake_case__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A , "do_resize" ) ) self.assertTrue(hasattr(__A , "size" ) ) self.assertTrue(hasattr(__A , "do_center_crop" ) ) self.assertTrue(hasattr(__A , "center_crop" ) ) self.assertTrue(hasattr(__A , "do_normalize" ) ) self.assertTrue(hasattr(__A , "image_mean" ) ) self.assertTrue(hasattr(__A , "image_std" ) ) self.assertTrue(hasattr(__A , "do_convert_rgb" ) ) def _lowercase ( self : Tuple ): snake_case__ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 2_2_4, "width": 2_2_4} ) self.assertEqual(image_processor.crop_size , {"height": 1_8, "width": 1_8} ) snake_case__ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {"shortest_edge": 4_2} ) self.assertEqual(image_processor.crop_size , {"height": 8_4, "width": 8_4} ) def _lowercase ( self : Tuple ): pass def _lowercase ( self : Union[str, Any] ): # Initialize image_processing snake_case__ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case__ : Optional[int] = self.image_processor_tester.prepare_inputs(equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A , Image.Image ) # Test not batched input snake_case__ : List[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case__ : Optional[int] = image_processing(__A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def _lowercase ( self : Dict ): # Initialize image_processing snake_case__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case__ : List[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A , numpify=__A ) for image in image_inputs: self.assertIsInstance(__A , np.ndarray ) # Test not batched input snake_case__ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case__ : List[str] = image_processing(__A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def _lowercase ( self : Union[str, Any] ): # Initialize image_processing snake_case__ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case__ : List[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A , torchify=__A ) for image in image_inputs: self.assertIsInstance(__A , torch.Tensor ) # Test not batched input snake_case__ : Dict = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case__ : Dict = image_processing(__A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( __UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = ChineseCLIPImageProcessor if is_vision_available() else None def _lowercase ( self : Optional[Any] ): snake_case__ : Tuple = ChineseCLIPImageProcessingTester(self , num_channels=4 , do_center_crop=__A ) snake_case__ : Union[str, Any] = 3 @property def _lowercase ( self : Dict ): return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : int ): snake_case__ : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A , "do_resize" ) ) self.assertTrue(hasattr(__A , "size" ) ) self.assertTrue(hasattr(__A , "do_center_crop" ) ) self.assertTrue(hasattr(__A , "center_crop" ) ) self.assertTrue(hasattr(__A , "do_normalize" ) ) self.assertTrue(hasattr(__A , "image_mean" ) ) self.assertTrue(hasattr(__A , "image_std" ) ) self.assertTrue(hasattr(__A , "do_convert_rgb" ) ) def _lowercase ( self : Optional[int] ): pass def _lowercase ( self : Any ): # Initialize image_processing snake_case__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case__ : Union[str, Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A , Image.Image ) # Test not batched input snake_case__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case__ : Optional[int] = image_processing(__A , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
297
from itertools import count def UpperCamelCase__ ( _A: int = 50 ): '''simple docstring''' __lowerCamelCase = [1] * min_block_length for n in count(_A ): fill_count_functions.append(1 ) for block_length in range(_A , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 1000000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
479
0
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments lowercase_ = logging.getLogger(__name__) @dataclass class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCAmelCase : Optional[float] = field( default=0.0 , metadata={'help': 'The label smoothing epsilon to apply (if not zero).'} ) __UpperCAmelCase : bool = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Whether to SortishSamler or not.'} ) __UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) __UpperCAmelCase : bool = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'whether to use adafactor'} ) __UpperCAmelCase : Optional[float] = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Encoder layer dropout probability. Goes into model.config.'} ) __UpperCAmelCase : Optional[float] = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Decoder layer dropout probability. Goes into model.config.'} ) __UpperCAmelCase : Optional[float] = field(default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Dropout probability. Goes into model.config.'} ) __UpperCAmelCase : Optional[float] = field( default=__SCREAMING_SNAKE_CASE , metadata={'help': 'Attention dropout probability. Goes into model.config.'} ) __UpperCAmelCase : Optional[str] = field( default='linear' , metadata={'help': f'Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}'} , )
65
"""simple docstring""" import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def lowercase ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[str] ) -> Any: if isinstance(lowerCAmelCase__ , torch.Tensor ): return image elif isinstance(lowerCAmelCase__ , PIL.Image.Image ): __a = [image] if isinstance(image[0] , PIL.Image.Image ): __a = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] __a = np.concatenate(lowerCAmelCase__ , axis=0 ) __a = np.array(lowerCAmelCase__ ).astype(np.floataa ) / 2_55.0 __a = image.transpose(0 , 3 , 1 , 2 ) __a = 2.0 * image - 1.0 __a = torch.from_numpy(lowerCAmelCase__ ) elif isinstance(image[0] , torch.Tensor ): __a = torch.cat(lowerCAmelCase__ , dim=0 ) return image def lowercase ( lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Optional[int]=0.99_95 ) -> int: if not isinstance(lowerCAmelCase__ , np.ndarray ): __a = True __a = va.device __a = va.cpu().numpy() __a = va.cpu().numpy() __a = np.sum(va * va / (np.linalg.norm(lowerCAmelCase__ ) * np.linalg.norm(lowerCAmelCase__ )) ) if np.abs(lowerCAmelCase__ ) > DOT_THRESHOLD: __a = (1 - t) * va + t * va else: __a = np.arccos(lowerCAmelCase__ ) __a = np.sin(lowerCAmelCase__ ) __a = theta_a * t __a = np.sin(lowerCAmelCase__ ) __a = np.sin(theta_a - theta_t ) / sin_theta_a __a = sin_theta_t / sin_theta_a __a = sa * va + sa * va if inputs_are_torch: __a = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) return va def lowercase ( lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] ) -> int: __a = F.normalize(lowerCAmelCase__ , dim=-1 ) __a = F.normalize(lowerCAmelCase__ , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def lowercase ( lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Any ) -> List[str]: for param in model.parameters(): __a = value class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , _a , _a , _a , _a , _a , _a , _a , _a=None , _a=None , _a=None , ): super().__init__() self.register_modules( vae=_a , text_encoder=_a , clip_model=_a , tokenizer=_a , unet=_a , scheduler=_a , feature_extractor=_a , coca_model=_a , coca_tokenizer=_a , coca_transform=_a , ) __a = ( feature_extractor.size if isinstance(feature_extractor.size , _a ) else feature_extractor.size['''shortest_edge'''] ) __a = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , _a ) set_requires_grad(self.clip_model , _a ) def __UpperCAmelCase ( self , _a = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __a = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_a ) def __UpperCAmelCase ( self ): self.enable_attention_slicing(_a ) def __UpperCAmelCase ( self ): set_requires_grad(self.vae , _a ) def __UpperCAmelCase ( self ): set_requires_grad(self.vae , _a ) def __UpperCAmelCase ( self ): set_requires_grad(self.unet , _a ) def __UpperCAmelCase ( self ): set_requires_grad(self.unet , _a ) def __UpperCAmelCase ( self , _a , _a , _a ): # get the original timestep using init_timestep __a = min(int(num_inference_steps * strength ) , _a ) __a = max(num_inference_steps - init_timestep , 0 ) __a = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a=None ): if not isinstance(_a , torch.Tensor ): raise ValueError(f'''`image` has to be of type `torch.Tensor` but is {type(_a )}''' ) __a = image.to(device=_a , dtype=_a ) if isinstance(_a , _a ): __a = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_a ) ] __a = torch.cat(_a , dim=0 ) else: __a = self.vae.encode(_a ).latent_dist.sample(_a ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __a = 0.1_8215 * init_latents __a = init_latents.repeat_interleave(_a , dim=0 ) __a = randn_tensor(init_latents.shape , generator=_a , device=_a , dtype=_a ) # get latents __a = self.scheduler.add_noise(_a , _a , _a ) __a = init_latents return latents def __UpperCAmelCase ( self , _a ): __a = self.coca_transform(_a ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): __a = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) __a = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split('''<end_of_text>''' )[0].replace('''<start_of_text>''' , '''''' ).rstrip(''' .,''' ) def __UpperCAmelCase ( self , _a , _a ): __a = self.feature_extractor.preprocess(_a ) __a = torch.from_numpy(clip_image_input['''pixel_values'''][0] ).unsqueeze(0 ).to(self.device ).half() __a = self.clip_model.get_image_features(_a ) __a = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=_a ) __a = image_embeddings_clip.repeat_interleave(_a , dim=0 ) return image_embeddings_clip @torch.enable_grad() def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a , _a , ): __a = latents.detach().requires_grad_() __a = self.scheduler.scale_model_input(_a , _a ) # predict the noise residual __a = self.unet(_a , _a , encoder_hidden_states=_a ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): __a = self.scheduler.alphas_cumprod[timestep] __a = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __a = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 __a = torch.sqrt(_a ) __a = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , _a ): __a = self.scheduler.sigmas[index] __a = latents - sigma * noise_pred else: raise ValueError(f'''scheduler type {type(self.scheduler )} not supported''' ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __a = 1 / 0.1_8215 * sample __a = self.vae.decode(_a ).sample __a = (image / 2 + 0.5).clamp(0 , 1 ) __a = transforms.Resize(self.feature_extractor_size )(_a ) __a = self.normalize(_a ).to(latents.dtype ) __a = self.clip_model.get_image_features(_a ) __a = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=_a ) __a = spherical_dist_loss(_a , _a ).mean() * clip_guidance_scale __a = -torch.autograd.grad(_a , _a )[0] if isinstance(self.scheduler , _a ): __a = latents.detach() + grads * (sigma**2) __a = noise_pred_original else: __a = noise_pred_original - torch.sqrt(_a ) * grads return noise_pred, latents @torch.no_grad() def __call__( self , _a , _a , _a = None , _a = None , _a = 512 , _a = 512 , _a = 0.6 , _a = 50 , _a = 7.5 , _a = 1 , _a = 0.0 , _a = 100 , _a = None , _a = "pil" , _a = True , _a = 0.8 , _a = 0.1 , _a = 0.1 , ): if isinstance(_a , _a ) and len(_a ) != batch_size: raise ValueError(f'''You have passed {batch_size} batch_size, but only {len(_a )} generators.''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if isinstance(_a , torch.Generator ) and batch_size > 1: __a = [generator] + [None] * (batch_size - 1) __a = [ ('''model''', self.coca_model is None), ('''tokenizer''', self.coca_tokenizer is None), ('''transform''', self.coca_transform is None), ] __a = [x[0] for x in coca_is_none if x[1]] __a = ''', '''.join(_a ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(_a ): raise ValueError( f'''Content prompt is None and CoCa [{coca_is_none_str}] is None.''' f'''Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) __a = self.get_image_description(_a ) if style_prompt is None: if len(_a ): raise ValueError( f'''Style prompt is None and CoCa [{coca_is_none_str}] is None.''' f''' Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) __a = self.get_image_description(_a ) # get prompt text embeddings for content and style __a = self.tokenizer( _a , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=_a , return_tensors='''pt''' , ) __a = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] __a = self.tokenizer( _a , padding='''max_length''' , max_length=self.tokenizer.model_max_length , truncation=_a , return_tensors='''pt''' , ) __a = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] __a = slerp(_a , _a , _a ) # duplicate text embeddings for each generation per prompt __a = text_embeddings.repeat_interleave(_a , dim=0 ) # set timesteps __a = '''offset''' in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) __a = {} if accepts_offset: __a = 1 self.scheduler.set_timesteps(_a , **_a ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) __a , __a = self.get_timesteps(_a , _a , self.device ) __a = timesteps[:1].repeat(_a ) # Preprocess image __a = preprocess(_a , _a , _a ) __a = self.prepare_latents( _a , _a , _a , text_embeddings.dtype , self.device , _a ) __a = preprocess(_a , _a , _a ) __a = self.prepare_latents( _a , _a , _a , text_embeddings.dtype , self.device , _a ) __a = slerp(_a , _a , _a ) if clip_guidance_scale > 0: __a = self.get_clip_image_embeddings(_a , _a ) __a = self.get_clip_image_embeddings(_a , _a ) __a = slerp( _a , _a , _a ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. __a = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: __a = content_text_input.input_ids.shape[-1] __a = self.tokenizer([''''''] , padding='''max_length''' , max_length=_a , return_tensors='''pt''' ) __a = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt __a = uncond_embeddings.repeat_interleave(_a , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __a = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. __a = (batch_size, self.unet.config.in_channels, height // 8, width // 8) __a = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps __a = torch.randn(_a , generator=_a , device='''cpu''' , dtype=_a ).to( self.device ) else: __a = torch.randn(_a , generator=_a , device=self.device , dtype=_a ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) __a = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __a = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __a = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __a = {} if accepts_eta: __a = eta # check if the scheduler accepts generator __a = '''generator''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: __a = generator with self.progress_bar(total=_a ): for i, t in enumerate(_a ): # expand the latents if we are doing classifier free guidance __a = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __a = self.scheduler.scale_model_input(_a , _a ) # predict the noise residual __a = self.unet(_a , _a , encoder_hidden_states=_a ).sample # perform classifier free guidance if do_classifier_free_guidance: __a , __a = noise_pred.chunk(2 ) __a = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: __a = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) __a , __a = self.cond_fn( _a , _a , _a , _a , _a , _a , _a , ) # compute the previous noisy sample x_t -> x_t-1 __a = self.scheduler.step(_a , _a , _a , **_a ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __a = 1 / 0.1_8215 * latents __a = self.vae.decode(_a ).sample __a = (image / 2 + 0.5).clamp(0 , 1 ) __a = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __a = self.numpy_to_pil(_a ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=_a , nsfw_content_detected=_a )
65
1
"""simple docstring""" import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) snake_case = '''\ Text data. Second line of data.''' snake_case = '''file''' @pytest.fixture(scope='''session''' ) def snake_case ( lowerCAmelCase_ ) -> Any: _snake_case = tmp_path_factory.mktemp('''data''' ) / (FILE_PATH + '''.zstd''') _snake_case = bytes(lowerCAmelCase_ , '''utf-8''' ) with zstd.open(lowerCAmelCase_ , '''wb''' ) as f: f.write(lowerCAmelCase_ ) return path @pytest.fixture def snake_case ( lowerCAmelCase_ ) -> Tuple: with open(os.path.join(tmpfs.local_root_dir , lowerCAmelCase_ ) , '''w''' ) as f: f.write(lowerCAmelCase_ ) return FILE_PATH @pytest.mark.parametrize('''compression_format''' , ['''gzip''', '''xz''', '''zstd'''] ) def snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Tuple: _snake_case = {'''gzip''': gz_file, '''xz''': xz_file, '''zstd''': zstd_path} _snake_case = input_paths[compression_format] _snake_case = tmp_path / '''cache''' _snake_case = DownloadConfig(cache_dir=lowerCAmelCase_ , extract_compressed_file=lowerCAmelCase_ ) _snake_case = cached_path(lowerCAmelCase_ , download_config=lowerCAmelCase_ ) with open(lowerCAmelCase_ ) as f: _snake_case = f.read() with open(lowerCAmelCase_ ) as f: _snake_case = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('''default_extracted''' , [True, False] ) @pytest.mark.parametrize('''default_cache_dir''' , [True, False] ) def snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Union[str, Any]: _snake_case = '''custom_cache''' _snake_case = '''custom_extracted_dir''' _snake_case = tmp_path / '''custom_extracted_path''' if default_extracted: _snake_case = ('''downloads''' if default_cache_dir else custom_cache_dir, '''extracted''') else: monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_DIR''' , lowerCAmelCase_ ) monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(lowerCAmelCase_ ) ) _snake_case = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) _snake_case = xz_file _snake_case = ( DownloadConfig(extract_compressed_file=lowerCAmelCase_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=lowerCAmelCase_ ) ) _snake_case = cached_path(lowerCAmelCase_ , download_config=lowerCAmelCase_ ) assert Path(lowerCAmelCase_ ).parent.parts[-2:] == expected def snake_case ( lowerCAmelCase_ ) -> List[str]: # absolute path _snake_case = str(Path(lowerCAmelCase_ ).resolve() ) assert cached_path(lowerCAmelCase_ ) == text_file # relative path _snake_case = str(Path(lowerCAmelCase_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowerCAmelCase_ ) == text_file def snake_case ( lowerCAmelCase_ ) -> Dict: # absolute path _snake_case = str(tmp_path.resolve() / '''__missing_file__.txt''' ) with pytest.raises(lowerCAmelCase_ ): cached_path(lowerCAmelCase_ ) # relative path _snake_case = '''./__missing_file__.txt''' with pytest.raises(lowerCAmelCase_ ): cached_path(lowerCAmelCase_ ) def snake_case ( lowerCAmelCase_ ) -> Any: _snake_case = get_from_cache(f"""tmp://{tmpfs_file}""" ) with open(lowerCAmelCase_ ) as f: _snake_case = f.read() assert output_file_content == FILE_CONTENT @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def snake_case ( ) -> Optional[Any]: with pytest.raises(lowerCAmelCase_ ): cached_path('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def snake_case ( lowerCAmelCase_ ) -> Optional[int]: _snake_case = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowerCAmelCase_ ): http_get('''https://huggingface.co''' , temp_file=lowerCAmelCase_ ) with pytest.raises(lowerCAmelCase_ ): http_head('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def snake_case ( lowerCAmelCase_ ) -> int: _snake_case = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowerCAmelCase_ ): ftp_get('''ftp://huggingface.co''' , temp_file=lowerCAmelCase_ ) with pytest.raises(lowerCAmelCase_ ): ftp_head('''ftp://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , lowerCAmelCase_ ) def snake_case ( lowerCAmelCase_ ) -> str: _snake_case = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(lowerCAmelCase_ ): fsspec_get('''s3://huggingface.co''' , temp_file=lowerCAmelCase_ ) with pytest.raises(lowerCAmelCase_ ): fsspec_head('''s3://huggingface.co''' )
103
"""simple docstring""" import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset UpperCAmelCase : List[Any] = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Dict , lowerCAmelCase_ : Union[str, Any]): """simple docstring""" super().__init__() lowercase_ = torchvision.models.resnetaaa(pretrained=lowerCAmelCase_) lowercase_ = list(model.children())[:-2] lowercase_ = nn.Sequential(*lowerCAmelCase_) lowercase_ = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds]) def _UpperCAmelCase ( self : Optional[int] , lowerCAmelCase_ : int): """simple docstring""" lowercase_ = self.pool(self.model(lowerCAmelCase_)) lowercase_ = torch.flatten(lowerCAmelCase_ , start_dim=2) lowercase_ = out.transpose(1 , 2).contiguous() return out # BxNx2048 class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): def __init__( self : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str]): """simple docstring""" lowercase_ = [json.loads(lowerCAmelCase_) for l in open(lowerCAmelCase_)] lowercase_ = os.path.dirname(lowerCAmelCase_) lowercase_ = tokenizer lowercase_ = labels lowercase_ = len(lowerCAmelCase_) lowercase_ = max_seq_length lowercase_ = transforms def __len__( self : List[Any]): """simple docstring""" return len(self.data) def __getitem__( self : Tuple , lowerCAmelCase_ : Union[str, Any]): """simple docstring""" lowercase_ = torch.LongTensor(self.tokenizer.encode(self.data[index]["""text"""] , add_special_tokens=lowerCAmelCase_)) lowercase_ , lowercase_ , lowercase_ = sentence[0], sentence[1:-1], sentence[-1] lowercase_ = sentence[: self.max_seq_length] lowercase_ = torch.zeros(self.n_classes) lowercase_ = 1 lowercase_ = Image.open(os.path.join(self.data_dir , self.data[index]["""img"""])).convert("""RGB""") lowercase_ = self.transforms(lowerCAmelCase_) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" lowercase_ = Counter() for row in self.data: label_freqs.update(row["""label"""]) return label_freqs def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Tuple: '''simple docstring''' lowercase_ = [len(row["""sentence"""] ) for row in batch] lowercase_ , lowercase_ = len(__lowerCAmelCase ), max(__lowerCAmelCase ) lowercase_ = torch.zeros(__lowerCAmelCase , __lowerCAmelCase , dtype=torch.long ) lowercase_ = torch.zeros(__lowerCAmelCase , __lowerCAmelCase , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(__lowerCAmelCase , __lowerCAmelCase ) ): lowercase_ = input_row["""sentence"""] lowercase_ = 1 lowercase_ = torch.stack([row["""image"""] for row in batch] ) lowercase_ = torch.stack([row["""label"""] for row in batch] ) lowercase_ = torch.stack([row["""image_start_token"""] for row in batch] ) lowercase_ = torch.stack([row["""image_end_token"""] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def _SCREAMING_SNAKE_CASE () -> Dict: '''simple docstring''' return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def _SCREAMING_SNAKE_CASE () -> List[Any]: '''simple docstring''' return transforms.Compose( [ transforms.Resize(2_56 ), transforms.CenterCrop(2_24 ), transforms.ToTensor(), transforms.Normalize( mean=[0.46_777_044, 0.44_531_429, 0.40_661_017] , std=[0.12_221_994, 0.12_145_835, 0.14_380_469] , ), ] )
567
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCAmelCase : str = { 'configuration_transfo_xl': ['TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TransfoXLConfig'], 'tokenization_transfo_xl': ['TransfoXLCorpus', 'TransfoXLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : Any = [ 'TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'AdaptiveEmbedding', 'TransfoXLForSequenceClassification', 'TransfoXLLMHeadModel', 'TransfoXLModel', 'TransfoXLPreTrainedModel', 'load_tf_weights_in_transfo_xl', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : List[Any] = [ 'TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFAdaptiveEmbedding', 'TFTransfoXLForSequenceClassification', 'TFTransfoXLLMHeadModel', 'TFTransfoXLMainLayer', 'TFTransfoXLModel', 'TFTransfoXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys __UpperCAmelCase : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
700
'''simple docstring''' import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename __UpperCAmelCase = 'http://www.mocksite.com/file1.txt' __UpperCAmelCase = '"text": ["foo", "foo"]' __UpperCAmelCase = '6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8' class _a : """simple docstring""" A = 2_00 A = {'Content-Length': '100'} A = {} def __a ( self ,**__SCREAMING_SNAKE_CASE ): return [bytes(__SCREAMING_SNAKE_CASE ,'utf-8' )] def SCREAMING_SNAKE_CASE_ ( *snake_case_ : Optional[Any] , **snake_case_ : Optional[int] ) -> str: return MockResponse() @pytest.mark.parametrize('urls_type' , [str, list, dict] ) def SCREAMING_SNAKE_CASE_ ( snake_case_ : List[str] , snake_case_ : List[Any] , snake_case_ : int ) -> Any: import requests monkeypatch.setattr(snake_case_ , 'request' , snake_case_ ) SCREAMING_SNAKE_CASE : int = URL if issubclass(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : Any = url elif issubclass(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : Union[str, Any] = [url] elif issubclass(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : List[Any] = {'train': url} SCREAMING_SNAKE_CASE : List[Any] = 'dummy' SCREAMING_SNAKE_CASE : str = 'downloads' SCREAMING_SNAKE_CASE : str = tmp_path SCREAMING_SNAKE_CASE : int = DownloadConfig( cache_dir=os.path.join(snake_case_ , snake_case_ ) , use_etag=snake_case_ , ) SCREAMING_SNAKE_CASE : Tuple = DownloadManager(dataset_name=snake_case_ , download_config=snake_case_ ) SCREAMING_SNAKE_CASE : Any = dl_manager.download(snake_case_ ) SCREAMING_SNAKE_CASE : List[str] = urls for downloaded_paths in [downloaded_paths]: if isinstance(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : List[str] = [downloaded_paths] SCREAMING_SNAKE_CASE : List[str] = [urls] elif isinstance(snake_case_ , snake_case_ ): assert "train" in downloaded_paths.keys() SCREAMING_SNAKE_CASE : Optional[Any] = downloaded_paths.values() SCREAMING_SNAKE_CASE : Dict = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(snake_case_ , snake_case_ ): assert downloaded_path == dl_manager.downloaded_paths[input_url] SCREAMING_SNAKE_CASE : List[Any] = Path(snake_case_ ) SCREAMING_SNAKE_CASE : List[Any] = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() SCREAMING_SNAKE_CASE : List[Any] = downloaded_path.read_text() assert content == CONTENT SCREAMING_SNAKE_CASE : int = downloaded_path.with_suffix('.json' ) assert metadata_downloaded_path.exists() SCREAMING_SNAKE_CASE : Union[str, Any] = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize('paths_type' , [str, list, dict] ) def SCREAMING_SNAKE_CASE_ ( snake_case_ : Optional[Any] , snake_case_ : Dict , snake_case_ : Tuple ) -> Any: SCREAMING_SNAKE_CASE : int = str(snake_case_ ) if issubclass(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : Dict = filename elif issubclass(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : List[str] = [filename] elif issubclass(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : Any = {'train': filename} SCREAMING_SNAKE_CASE : Dict = 'dummy' SCREAMING_SNAKE_CASE : Tuple = xz_file.parent SCREAMING_SNAKE_CASE : Any = 'extracted' SCREAMING_SNAKE_CASE : Optional[int] = DownloadConfig( cache_dir=snake_case_ , use_etag=snake_case_ , ) SCREAMING_SNAKE_CASE : List[str] = DownloadManager(dataset_name=snake_case_ , download_config=snake_case_ ) SCREAMING_SNAKE_CASE : List[str] = dl_manager.extract(snake_case_ ) SCREAMING_SNAKE_CASE : Any = paths for extracted_paths in [extracted_paths]: if isinstance(snake_case_ , snake_case_ ): SCREAMING_SNAKE_CASE : Any = [extracted_paths] SCREAMING_SNAKE_CASE : Tuple = [paths] elif isinstance(snake_case_ , snake_case_ ): assert "train" in extracted_paths.keys() SCREAMING_SNAKE_CASE : Dict = extracted_paths.values() SCREAMING_SNAKE_CASE : Union[str, Any] = paths.values() assert extracted_paths for extracted_path, input_path in zip(snake_case_ , snake_case_ ): assert extracted_path == dl_manager.extracted_paths[input_path] SCREAMING_SNAKE_CASE : Optional[int] = Path(snake_case_ ) SCREAMING_SNAKE_CASE : int = extracted_path.parts assert parts[-1] == hash_url_to_filename(snake_case_ , etag=snake_case_ ) assert parts[-2] == extracted_subdir assert extracted_path.exists() SCREAMING_SNAKE_CASE : List[Any] = extracted_path.read_text() SCREAMING_SNAKE_CASE : Any = text_file.read_text() assert extracted_file_content == expected_file_content def SCREAMING_SNAKE_CASE_ ( snake_case_ : Union[str, Any] , snake_case_ : Optional[int] ) -> List[str]: assert path.endswith('.jsonl' ) for num_items, line in enumerate(snake_case_ , start=1 ): SCREAMING_SNAKE_CASE : Optional[int] = json.loads(line.decode('utf-8' ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize('archive_jsonl' , ['tar_jsonl_path', 'zip_jsonl_path'] ) def SCREAMING_SNAKE_CASE_ ( snake_case_ : Tuple , snake_case_ : Optional[int] ) -> Dict: SCREAMING_SNAKE_CASE : int = request.getfixturevalue(snake_case_ ) SCREAMING_SNAKE_CASE : Optional[int] = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(snake_case_ ) , start=1 ): _test_jsonl(snake_case_ , snake_case_ ) assert num_jsonl == 2 @pytest.mark.parametrize('archive_nested_jsonl' , ['tar_nested_jsonl_path', 'zip_nested_jsonl_path'] ) def SCREAMING_SNAKE_CASE_ ( snake_case_ : List[str] , snake_case_ : Optional[Any] ) -> List[str]: SCREAMING_SNAKE_CASE : str = request.getfixturevalue(snake_case_ ) SCREAMING_SNAKE_CASE : int = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(snake_case_ ) , start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(snake_case_ ) , start=1 ): _test_jsonl(snake_case_ , snake_case_ ) assert num_tar == 1 assert num_jsonl == 2 def SCREAMING_SNAKE_CASE_ ( snake_case_ : Optional[Any] ) -> List[Any]: SCREAMING_SNAKE_CASE : Optional[Any] = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(snake_case_ ) , start=1 ): assert os.path.basename(snake_case_ ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
220
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'openai-gpt': 'https://huggingface.co/openai-gpt/resolve/main/config.json'} class lowerCAmelCase__ ( a): '''simple docstring''' __SCREAMING_SNAKE_CASE = "openai-gpt" __SCREAMING_SNAKE_CASE = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , __lowerCamelCase=4_0_4_7_8 , __lowerCamelCase=5_1_2 , __lowerCamelCase=7_6_8 , __lowerCamelCase=1_2 , __lowerCamelCase=1_2 , __lowerCamelCase="gelu" , __lowerCamelCase=0.1 , __lowerCamelCase=0.1 , __lowerCamelCase=0.1 , __lowerCamelCase=1e-5 , __lowerCamelCase=0.0_2 , __lowerCamelCase="cls_index" , __lowerCamelCase=True , __lowerCamelCase=None , __lowerCamelCase=True , __lowerCamelCase=0.1 , **__lowerCamelCase , ) -> str: _A : Dict = vocab_size _A : Optional[Any] = n_positions _A : List[Any] = n_embd _A : int = n_layer _A : Optional[Any] = n_head _A : List[str] = afn _A : int = resid_pdrop _A : int = embd_pdrop _A : Any = attn_pdrop _A : Tuple = layer_norm_epsilon _A : Optional[Any] = initializer_range _A : Optional[Any] = summary_type _A : Optional[int] = summary_use_proj _A : Dict = summary_activation _A : List[str] = summary_first_dropout _A : Dict = summary_proj_to_labels super().__init__(**__lowerCamelCase)
503
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase__ ( a , unittest.TestCase): '''simple docstring''' __SCREAMING_SNAKE_CASE = KandinskyVaaControlnetImgaImgPipeline __SCREAMING_SNAKE_CASE = ["image_embeds", "negative_image_embeds", "image", "hint"] __SCREAMING_SNAKE_CASE = ["image_embeds", "negative_image_embeds", "image", "hint"] __SCREAMING_SNAKE_CASE = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] __SCREAMING_SNAKE_CASE = False @property def _lowerCamelCase ( self) -> Tuple: return 3_2 @property def _lowerCamelCase ( self) -> Dict: return 3_2 @property def _lowerCamelCase ( self) -> str: return self.time_input_dim @property def _lowerCamelCase ( self) -> str: return self.time_input_dim * 4 @property def _lowerCamelCase ( self) -> str: return 1_0_0 @property def _lowerCamelCase ( self) -> Tuple: torch.manual_seed(0) _A : str = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _A : str = UNetaDConditionModel(**__lowerCamelCase) return model @property def _lowerCamelCase ( self) -> str: return { "block_out_channels": [3_2, 3_2, 6_4, 6_4], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def _lowerCamelCase ( self) -> List[Any]: torch.manual_seed(0) _A : int = VQModel(**self.dummy_movq_kwargs) return model def _lowerCamelCase ( self) -> Any: _A : List[str] = self.dummy_unet _A : Optional[int] = self.dummy_movq _A : str = { "num_train_timesteps": 1_0_0_0, "beta_schedule": "linear", "beta_start": 0.0_0_0_8_5, "beta_end": 0.0_1_2, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } _A : int = DDIMScheduler(**__lowerCamelCase) _A : Optional[Any] = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase=0) -> Any: _A : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__lowerCamelCase)).to(__lowerCamelCase) _A : List[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1)).to( __lowerCamelCase) # create init_image _A : List[Any] = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(__lowerCamelCase)).to(__lowerCamelCase) _A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1)[0] _A : Union[str, Any] = Image.fromarray(np.uinta(__lowerCamelCase)).convert("RGB").resize((2_5_6, 2_5_6)) # create hint _A : Dict = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(__lowerCamelCase)).to(__lowerCamelCase) if str(__lowerCamelCase).startswith("mps"): _A : Dict = torch.manual_seed(__lowerCamelCase) else: _A : Tuple = torch.Generator(device=__lowerCamelCase).manual_seed(__lowerCamelCase) _A : Tuple = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 6_4, "width": 6_4, "num_inference_steps": 1_0, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def _lowerCamelCase ( self) -> int: _A : List[Any] = "cpu" _A : List[Any] = self.get_dummy_components() _A : Any = self.pipeline_class(**__lowerCamelCase) _A : Any = pipe.to(__lowerCamelCase) pipe.set_progress_bar_config(disable=__lowerCamelCase) _A : str = pipe(**self.get_dummy_inputs(__lowerCamelCase)) _A : int = output.images _A : Dict = pipe( **self.get_dummy_inputs(__lowerCamelCase) , return_dict=__lowerCamelCase , )[0] _A : Any = image[0, -3:, -3:, -1] _A : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _A : List[str] = np.array( [0.5_4_9_8_5_0_3_4, 0.5_5_5_0_9_3_6_5, 0.5_2_5_6_1_5_0_4, 0.5_5_7_0_4_9_4, 0.5_5_9_3_8_1_8, 0.5_2_6_3_9_7_9, 0.5_0_2_8_5_6_4_3, 0.5_0_6_9_8_4_6, 0.5_1_1_9_6_7_3_6]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase): '''simple docstring''' def _lowerCamelCase ( self) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self) -> Tuple: _A : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy") _A : Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") _A : str = init_image.resize((5_1_2, 5_1_2)) _A : Optional[int] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png") _A : Optional[Any] = torch.from_numpy(np.array(__lowerCamelCase)).float() / 2_5_5.0 _A : Optional[Any] = hint.permute(2 , 0 , 1).unsqueeze(0) _A : List[str] = "A robot, 4k photo" _A : Any = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa) pipe_prior.to(__lowerCamelCase) _A : List[str] = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa) _A : Dict = pipeline.to(__lowerCamelCase) pipeline.set_progress_bar_config(disable=__lowerCamelCase) _A : Tuple = torch.Generator(device="cpu").manual_seed(0) _A , _A : List[str] = pipe_prior( __lowerCamelCase , image=__lowerCamelCase , strength=0.8_5 , generator=__lowerCamelCase , negative_prompt="" , ).to_tuple() _A : Optional[int] = pipeline( image=__lowerCamelCase , image_embeds=__lowerCamelCase , negative_image_embeds=__lowerCamelCase , hint=__lowerCamelCase , generator=__lowerCamelCase , num_inference_steps=1_0_0 , height=5_1_2 , width=5_1_2 , strength=0.5 , output_type="np" , ) _A : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert_mean_pixel_difference(__lowerCamelCase , __lowerCamelCase)
503
1
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def __snake_case ( _UpperCamelCase ) -> str: _a , _a = image.size _a , _a = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 _a = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) _a = np.array(_UpperCamelCase ).astype(np.floataa ) / 2_55.0 _a = image[None].transpose(0 , 3 , 1 , 2 ) _a = torch.from_numpy(_UpperCamelCase ) return 2.0 * image - 1.0 class UpperCAmelCase ( __snake_case ): def __init__( self: Dict , __UpperCamelCase: VQModel , __UpperCamelCase: UNetaDModel , __UpperCamelCase: Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): super().__init__() self.register_modules(vqvae=__UpperCamelCase , unet=__UpperCamelCase , scheduler=__UpperCamelCase ) @torch.no_grad() def __call__( self: int , __UpperCamelCase: Union[torch.Tensor, PIL.Image.Image] = None , __UpperCamelCase: Optional[int] = 1 , __UpperCamelCase: Optional[int] = 100 , __UpperCamelCase: Optional[float] = 0.0 , __UpperCamelCase: Optional[Union[torch.Generator, List[torch.Generator]]] = None , __UpperCamelCase: Optional[str] = "pil" , __UpperCamelCase: bool = True , ): if isinstance(__UpperCamelCase , PIL.Image.Image ): _a = 1 elif isinstance(__UpperCamelCase , torch.Tensor ): _a = image.shape[0] else: raise ValueError(f"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(__UpperCamelCase )}" ) if isinstance(__UpperCamelCase , PIL.Image.Image ): _a = preprocess(__UpperCamelCase ) _a , _a = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image _a = (batch_size, self.unet.config.in_channels // 2, height, width) _a = next(self.unet.parameters() ).dtype _a = randn_tensor(__UpperCamelCase , generator=__UpperCamelCase , device=self.device , dtype=__UpperCamelCase ) _a = image.to(device=self.device , dtype=__UpperCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) _a = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler _a = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] _a = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) _a = {} if accepts_eta: _a = eta for t in self.progress_bar(__UpperCamelCase ): # concat latents and low resolution image in the channel dimension. _a = torch.cat([latents, image] , dim=1 ) _a = self.scheduler.scale_model_input(__UpperCamelCase , __UpperCamelCase ) # predict the noise residual _a = self.unet(__UpperCamelCase , __UpperCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _a = self.scheduler.step(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ).prev_sample # decode the image latents with the VQVAE _a = self.vqvae.decode(__UpperCamelCase ).sample _a = torch.clamp(__UpperCamelCase , -1.0 , 1.0 ) _a = image / 2 + 0.5 _a = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _a = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase )
346
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class UpperCAmelCase : @staticmethod def _A ( *__UpperCamelCase: Optional[int] , **__UpperCamelCase: str ): pass def __snake_case ( _UpperCamelCase ) -> Dict: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. lowerCamelCase :List[str] = ( 'https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png' ) @is_pipeline_test @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): a: List[str] = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def _A ( self: Dict , __UpperCamelCase: Optional[int] , __UpperCamelCase: Tuple , __UpperCamelCase: Dict ): _a = pipeline( '''document-question-answering''' , model=__UpperCamelCase , tokenizer=__UpperCamelCase , image_processor=__UpperCamelCase ) _a = INVOICE_URL _a = list(zip(*apply_tesseract(load_image(__UpperCamelCase ) , __UpperCamelCase , '''''' ) ) ) _a = '''What is the placebo?''' _a = [ { '''image''': load_image(__UpperCamelCase ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def _A ( self: Tuple , __UpperCamelCase: Dict , __UpperCamelCase: List[str] ): _a = dqa_pipeline(__UpperCamelCase , top_k=2 ) self.assertEqual( __UpperCamelCase , [ [ {'''score''': ANY(__UpperCamelCase ), '''answer''': ANY(__UpperCamelCase ), '''start''': ANY(__UpperCamelCase ), '''end''': ANY(__UpperCamelCase )}, {'''score''': ANY(__UpperCamelCase ), '''answer''': ANY(__UpperCamelCase ), '''start''': ANY(__UpperCamelCase ), '''end''': ANY(__UpperCamelCase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def _A ( self: List[str] ): _a = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) _a = INVOICE_URL _a = '''How many cats are there?''' _a = [ {'''score''': 0.0_0_0_1, '''answer''': '''oy 2312/2019''', '''start''': 38, '''end''': 39}, {'''score''': 0.0_0_0_1, '''answer''': '''oy 2312/2019 DUE''', '''start''': 38, '''end''': 40}, ] _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCamelCase , decimals=4 ) , __UpperCamelCase ) _a = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__UpperCamelCase , decimals=4 ) , __UpperCamelCase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably _a = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual(__UpperCamelCase , [] ) # We can optionnally pass directly the words and bounding boxes _a = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _a = [] _a = [] _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , words=__UpperCamelCase , boxes=__UpperCamelCase , top_k=2 ) self.assertEqual(__UpperCamelCase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def _A ( self: Tuple ): _a = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) _a = INVOICE_URL _a = '''What is the invoice number?''' _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) _a = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) _a = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def _A ( self: Dict ): _a = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=50 , ) _a = INVOICE_URL _a = '''What is the invoice number?''' _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) _a = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) _a = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def _A ( self: Union[str, Any] ): _a = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCamelCase ) _a = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCamelCase , revision='''3dc6de3''' , ) _a = INVOICE_URL _a = '''What is the invoice number?''' _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) _a = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) _a = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] ] * 2 , ) _a = list(zip(*apply_tesseract(load_image(__UpperCamelCase ) , __UpperCamelCase , '''''' ) ) ) # This model should also work if `image` is set to None _a = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def _A ( self: List[Any] ): _a = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCamelCase ) _a = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCamelCase , revision='''3dc6de3''' , max_seq_len=50 , ) _a = INVOICE_URL _a = '''What is the invoice number?''' _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) _a = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) _a = list(zip(*apply_tesseract(load_image(__UpperCamelCase ) , __UpperCamelCase , '''''' ) ) ) # This model should also work if `image` is set to None _a = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCamelCase , decimals=4 ) , [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) @slow @require_torch def _A ( self: Optional[Any] ): _a = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) _a = INVOICE_URL _a = '''What is the invoice number?''' _a = dqa_pipeline(image=__UpperCamelCase , question=__UpperCamelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCamelCase , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def _A ( self: str ): pass
346
1
import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __A( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = jnp.floataa SCREAMING_SNAKE_CASE__ = True def UpperCAmelCase_ (self ): super().setup() UpperCamelCase__ = nn.Dense(5 , dtype=self.dtype ) def __call__(self , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = super().__call__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class __A( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ = FlaxBigBirdForNaturalQuestionsModule def __magic_name__ ( __a : str , __a : List[str] , __a : List[Any] , __a : Optional[int] , __a : int , __a : Tuple ): '''simple docstring''' def cross_entropy(__a : Optional[Any] , __a : Tuple , __a : List[Any]=None ): UpperCamelCase__ = logits.shape[-1] UpperCamelCase__ = (labels[..., None] == jnp.arange(__a )[None]).astype("""f4""" ) UpperCamelCase__ = jax.nn.log_softmax(__a , axis=-1 ) UpperCamelCase__ = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: UpperCamelCase__ = reduction(__a ) return loss UpperCamelCase__ = partial(__a , reduction=jnp.mean ) UpperCamelCase__ = cross_entropy(__a , __a ) UpperCamelCase__ = cross_entropy(__a , __a ) UpperCamelCase__ = cross_entropy(__a , __a ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __A: """simple docstring""" SCREAMING_SNAKE_CASE__ = "google/bigbird-roberta-base" SCREAMING_SNAKE_CASE__ = 3000 SCREAMING_SNAKE_CASE__ = 10500 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 3 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 5 # tx_args SCREAMING_SNAKE_CASE__ = 3e-5 SCREAMING_SNAKE_CASE__ = 0.0 SCREAMING_SNAKE_CASE__ = 20000 SCREAMING_SNAKE_CASE__ = 0.0095 SCREAMING_SNAKE_CASE__ = "bigbird-roberta-natural-questions" SCREAMING_SNAKE_CASE__ = "training-expt" SCREAMING_SNAKE_CASE__ = "data/nq-training.jsonl" SCREAMING_SNAKE_CASE__ = "data/nq-validation.jsonl" def UpperCAmelCase_ (self ): os.makedirs(self.base_dir , exist_ok=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = os.path.join(self.base_dir , self.save_dir ) UpperCamelCase__ = self.batch_size_per_device * jax.device_count() @dataclass class __A: """simple docstring""" SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 4096 # no dynamic padding on TPUs def __call__(self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = self.collate_fn(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = jax.tree_util.tree_map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return batch def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ , UpperCamelCase__ = self.fetch_inputs(features["""input_ids"""] ) UpperCamelCase__ = { """input_ids""": jnp.array(SCREAMING_SNAKE_CASE_ , dtype=jnp.intaa ), """attention_mask""": jnp.array(SCREAMING_SNAKE_CASE_ , dtype=jnp.intaa ), """start_labels""": jnp.array(features["""start_token"""] , dtype=jnp.intaa ), """end_labels""": jnp.array(features["""end_token"""] , dtype=jnp.intaa ), """pooled_labels""": jnp.array(features["""category"""] , dtype=jnp.intaa ), } return batch def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = [self._fetch_inputs(SCREAMING_SNAKE_CASE_ ) for ids in input_ids] return zip(*SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = [1 for _ in range(len(SCREAMING_SNAKE_CASE_ ) )] while len(SCREAMING_SNAKE_CASE_ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def __magic_name__ ( __a : List[Any] , __a : List[str] , __a : Tuple=None ): '''simple docstring''' if seed is not None: UpperCamelCase__ = dataset.shuffle(seed=__a ) for i in range(len(__a ) // batch_size ): UpperCamelCase__ = dataset[i * batch_size : (i + 1) * batch_size] yield dict(__a ) @partial(jax.pmap , axis_name="""batch""" ) def __magic_name__ ( __a : Dict , __a : List[Any] , **__a : int ): '''simple docstring''' def loss_fn(__a : Optional[Any] ): UpperCamelCase__ = model_inputs.pop("""start_labels""" ) UpperCamelCase__ = model_inputs.pop("""end_labels""" ) UpperCamelCase__ = model_inputs.pop("""pooled_labels""" ) UpperCamelCase__ = state.apply_fn(**__a , params=__a , dropout_rng=__a , train=__a ) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = outputs return state.loss_fn( __a , __a , __a , __a , __a , __a , ) UpperCamelCase__ , UpperCamelCase__ = jax.random.split(__a ) UpperCamelCase__ = jax.value_and_grad(__a ) UpperCamelCase__ , UpperCamelCase__ = grad_fn(state.params ) UpperCamelCase__ = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) UpperCamelCase__ = jax.lax.pmean(__a , """batch""" ) UpperCamelCase__ = state.apply_gradients(grads=__a ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="""batch""" ) def __magic_name__ ( __a : Optional[Any] , **__a : Any ): '''simple docstring''' UpperCamelCase__ = model_inputs.pop("""start_labels""" ) UpperCamelCase__ = model_inputs.pop("""end_labels""" ) UpperCamelCase__ = model_inputs.pop("""pooled_labels""" ) UpperCamelCase__ = state.apply_fn(**__a , params=state.params , train=__a ) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = outputs UpperCamelCase__ = state.loss_fn(__a , __a , __a , __a , __a , __a ) UpperCamelCase__ = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) return metrics class __A( train_state.TrainState ): """simple docstring""" SCREAMING_SNAKE_CASE__ = struct.field(pytree_node=__lowerCamelCase ) @dataclass class __A: """simple docstring""" SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = None def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None ): UpperCamelCase__ = model.params UpperCamelCase__ = TrainState.create( apply_fn=model.__call__ , params=SCREAMING_SNAKE_CASE_ , tx=SCREAMING_SNAKE_CASE_ , loss_fn=SCREAMING_SNAKE_CASE_ , ) if ckpt_dir is not None: UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = restore_checkpoint(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = { """lr""": args.lr, """init_lr""": args.init_lr, """warmup_steps""": args.warmup_steps, """num_train_steps""": num_train_steps, """weight_decay""": args.weight_decay, } UpperCamelCase__ , UpperCamelCase__ = build_tx(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = train_state.TrainState( step=SCREAMING_SNAKE_CASE_ , apply_fn=model.__call__ , params=SCREAMING_SNAKE_CASE_ , tx=SCREAMING_SNAKE_CASE_ , opt_state=SCREAMING_SNAKE_CASE_ , ) UpperCamelCase__ = args UpperCamelCase__ = data_collator UpperCamelCase__ = lr UpperCamelCase__ = params UpperCamelCase__ = jax_utils.replicate(SCREAMING_SNAKE_CASE_ ) return state def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = self.args UpperCamelCase__ = len(SCREAMING_SNAKE_CASE_ ) // args.batch_size UpperCamelCase__ = jax.random.PRNGKey(0 ) UpperCamelCase__ = jax.random.split(SCREAMING_SNAKE_CASE_ , jax.device_count() ) for epoch in range(args.max_epochs ): UpperCamelCase__ = jnp.array(0 , dtype=jnp.floataa ) UpperCamelCase__ = get_batched_dataset(SCREAMING_SNAKE_CASE_ , args.batch_size , seed=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = 0 for batch in tqdm(SCREAMING_SNAKE_CASE_ , total=SCREAMING_SNAKE_CASE_ , desc=F"Running EPOCH-{epoch}" ): UpperCamelCase__ = self.data_collator(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = self.train_step_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 if i % args.logging_steps == 0: UpperCamelCase__ = jax_utils.unreplicate(state.step ) UpperCamelCase__ = running_loss.item() / i UpperCamelCase__ = self.scheduler_fn(state_step - 1 ) UpperCamelCase__ = self.evaluate(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = { """step""": state_step.item(), """eval_loss""": eval_loss.item(), """tr_loss""": tr_loss, """lr""": lr.item(), } tqdm.write(str(SCREAMING_SNAKE_CASE_ ) ) self.logger.log(SCREAMING_SNAKE_CASE_ , commit=SCREAMING_SNAKE_CASE_ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + F"-e{epoch}-s{i}" , state=SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = get_batched_dataset(SCREAMING_SNAKE_CASE_ , self.args.batch_size ) UpperCamelCase__ = len(SCREAMING_SNAKE_CASE_ ) // self.args.batch_size UpperCamelCase__ = jnp.array(0 , dtype=jnp.floataa ) UpperCamelCase__ = 0 for batch in tqdm(SCREAMING_SNAKE_CASE_ , total=SCREAMING_SNAKE_CASE_ , desc="""Evaluating ... """ ): UpperCamelCase__ = self.data_collator(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = self.val_step_fn(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 return running_loss / i def UpperCAmelCase_ (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase__ = jax_utils.unreplicate(SCREAMING_SNAKE_CASE_ ) print(F"SAVING CHECKPOINT IN {save_dir}" , end=""" ... """ ) self.model_save_fn(SCREAMING_SNAKE_CASE_ , params=state.params ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , """opt_state.msgpack""" ) , """wb""" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(SCREAMING_SNAKE_CASE_ , """args.joblib""" ) ) joblib.dump(self.data_collator , os.path.join(SCREAMING_SNAKE_CASE_ , """data_collator.joblib""" ) ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , """training_state.json""" ) , """w""" ) as f: json.dump({"""step""": state.step.item()} , SCREAMING_SNAKE_CASE_ ) print("""DONE""" ) def __magic_name__ ( __a : str , __a : Union[str, Any] ): '''simple docstring''' print(f"RESTORING CHECKPOINT FROM {save_dir}" , end=""" ... """ ) with open(os.path.join(__a , """flax_model.msgpack""" ) , """rb""" ) as f: UpperCamelCase__ = from_bytes(state.params , f.read() ) with open(os.path.join(__a , """opt_state.msgpack""" ) , """rb""" ) as f: UpperCamelCase__ = from_bytes(state.opt_state , f.read() ) UpperCamelCase__ = joblib.load(os.path.join(__a , """args.joblib""" ) ) UpperCamelCase__ = joblib.load(os.path.join(__a , """data_collator.joblib""" ) ) with open(os.path.join(__a , """training_state.json""" ) , """r""" ) as f: UpperCamelCase__ = json.load(__a ) UpperCamelCase__ = training_state["""step"""] print("""DONE""" ) return params, opt_state, step, args, data_collator def __magic_name__ ( __a : Union[str, Any] , __a : List[Any] , __a : Optional[Any] , __a : Dict ): '''simple docstring''' UpperCamelCase__ = num_train_steps - warmup_steps UpperCamelCase__ = optax.linear_schedule(init_value=__a , end_value=__a , transition_steps=__a ) UpperCamelCase__ = optax.linear_schedule(init_value=__a , end_value=1E-7 , transition_steps=__a ) UpperCamelCase__ = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __magic_name__ ( __a : Dict , __a : List[str] , __a : Optional[Any] , __a : List[str] , __a : List[Any] ): '''simple docstring''' def weight_decay_mask(__a : List[Any] ): UpperCamelCase__ = traverse_util.flatten_dict(__a ) UpperCamelCase__ = {k: (v[-1] != """bias""" and v[-2:] != ("""LayerNorm""", """scale""")) for k, v in params.items()} return traverse_util.unflatten_dict(__a ) UpperCamelCase__ = scheduler_fn(__a , __a , __a , __a ) UpperCamelCase__ = optax.adamw(learning_rate=__a , weight_decay=__a , mask=__a ) return tx, lr
513
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def __magic_name__ ( __a : int , __a : Union[str, Any] , __a : Optional[int] , __a : Any ): '''simple docstring''' for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), f"Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})" else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), f"Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})" def __magic_name__ ( __a : List[Any] , __a : List[str] , __a : Optional[Any] , __a : Any , __a : Any=True ): '''simple docstring''' model.train() UpperCamelCase__ = model(__a ) UpperCamelCase__ = F.mse_loss(__a , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(__a ) def __magic_name__ ( __a : Any , __a : int=False ): '''simple docstring''' set_seed(42 ) UpperCamelCase__ = RegressionModel() UpperCamelCase__ = deepcopy(__a ) UpperCamelCase__ = RegressionDataset(length=80 ) UpperCamelCase__ = DataLoader(__a , batch_size=16 ) model.to(accelerator.device ) if sched: UpperCamelCase__ = AdamW(params=model.parameters() , lr=1E-3 ) UpperCamelCase__ = AdamW(params=ddp_model.parameters() , lr=1E-3 ) UpperCamelCase__ = LambdaLR(__a , lr_lambda=lambda __a : epoch**0.65 ) UpperCamelCase__ = LambdaLR(__a , lr_lambda=lambda __a : epoch**0.65 ) # Make a copy of `model` if sched: UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = accelerator.prepare(__a , __a , __a , __a ) else: UpperCamelCase__ , UpperCamelCase__ = accelerator.prepare(__a , __a ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def __magic_name__ ( __a : Optional[Any] ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = get_training_setup(__a ) # Use a single batch UpperCamelCase__ , UpperCamelCase__ = next(iter(__a ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model UpperCamelCase__ , UpperCamelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCamelCase__ , UpperCamelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(__a , __a , __a , __a ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(__a ): step_model(__a , __a , __a , __a ) else: # Sync grads step_model(__a , __a , __a , __a ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(__a , __a , __a , __a ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), f"Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})" # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) UpperCamelCase__ = ddp_input[torch.randperm(len(__a ) )] def __magic_name__ ( __a : Dict ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = get_training_setup(__a ) # Use a single batch UpperCamelCase__ , UpperCamelCase__ = next(iter(__a ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model UpperCamelCase__ , UpperCamelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCamelCase__ , UpperCamelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(__a , __a , __a , __a ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(__a ): step_model(__a , __a , __a , __a ) else: # Sync grads step_model(__a , __a , __a , __a ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), f"Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})" else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), f"Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})" # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) UpperCamelCase__ = ddp_input[torch.randperm(len(__a ) )] def __magic_name__ ( __a : List[Any]=False , __a : List[str]=False ): '''simple docstring''' UpperCamelCase__ = Accelerator( split_batches=__a , dispatch_batches=__a , gradient_accumulation_steps=2 ) # Test that context manager behaves properly UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = get_training_setup(__a ) for iteration, batch in enumerate(__a ): UpperCamelCase__ , UpperCamelCase__ = batch.values() # Gather the distributed inputs and targs for the base model UpperCamelCase__ , UpperCamelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCamelCase__ , UpperCamelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(__a , __a , __a , __a , __a ) # Do "gradient accumulation" (noop) with accelerator.accumulate(__a ): step_model(__a , __a , __a , __a ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(__a ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), f"Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})" else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), f"Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})" # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) UpperCamelCase__ = ddp_input[torch.randperm(len(__a ) )] GradientState._reset_state() def __magic_name__ ( __a : Optional[Any]=False , __a : int=False ): '''simple docstring''' UpperCamelCase__ = Accelerator( split_batches=__a , dispatch_batches=__a , gradient_accumulation_steps=2 ) # Test that context manager behaves properly UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = get_training_setup(__a , __a ) for iteration, batch in enumerate(__a ): UpperCamelCase__ , UpperCamelCase__ = batch.values() # Gather the distributed inputs and targs for the base model UpperCamelCase__ , UpperCamelCase__ = accelerator.gather((ddp_input, ddp_target) ) UpperCamelCase__ , UpperCamelCase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(__a , __a , __a , __a , __a ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(__a )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(__a ): step_model(__a , __a , __a , __a ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), f"Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n" UpperCamelCase__ = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(__a )) if accelerator.num_processes > 1: check_model_parameters(__a , __a , __a , __a ) # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) GradientState._reset_state() def __magic_name__ ( ): '''simple docstring''' UpperCamelCase__ = Accelerator() UpperCamelCase__ = RegressionDataset(length=80 ) UpperCamelCase__ = DataLoader(__a , batch_size=16 ) UpperCamelCase__ = RegressionDataset(length=96 ) UpperCamelCase__ = DataLoader(__a , batch_size=16 ) UpperCamelCase__ , UpperCamelCase__ = accelerator.prepare(__a , __a ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(__a ): assert id(accelerator.gradient_state.active_dataloader ) == id(__a ) if iteration < len(__a ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(__a ): assert id(accelerator.gradient_state.active_dataloader ) == id(__a ) if batch_num < len(__a ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def __magic_name__ ( ): '''simple docstring''' UpperCamelCase__ = Accelerator() UpperCamelCase__ = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(__a ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(__a ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """ , f"`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**" , ) test_gradient_accumulation(__a , __a ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""" , """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , """`split_batches=False`, `dispatch_batches=False`**""" , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , f"`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**" , ) test_gradient_accumulation_with_opt_and_scheduler(__a , __a ) def __magic_name__ ( __a : List[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
513
1
"""simple docstring""" from __future__ import annotations def A__ ( _UpperCAmelCase : int = 4 ) -> list[list[int]]: '''simple docstring''' snake_case__ : Any = abs(_UpperCAmelCase ) or 4 return [[1 + x + y * row_size for x in range(_UpperCAmelCase )] for y in range(_UpperCAmelCase )] def A__ ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: '''simple docstring''' return reverse_row(transpose(_UpperCAmelCase ) ) # OR.. transpose(reverse_column(matrix)) def A__ ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: '''simple docstring''' return reverse_row(reverse_column(_UpperCAmelCase ) ) # OR.. reverse_column(reverse_row(matrix)) def A__ ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: '''simple docstring''' return reverse_column(transpose(_UpperCAmelCase ) ) # OR.. transpose(reverse_row(matrix)) def A__ ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: '''simple docstring''' snake_case__ : Union[str, Any] = [list(_UpperCAmelCase ) for x in zip(*_UpperCAmelCase )] return matrix def A__ ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: '''simple docstring''' snake_case__ : Tuple = matrix[::-1] return matrix def A__ ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: '''simple docstring''' snake_case__ : Tuple = [x[::-1] for x in matrix] return matrix def A__ ( _UpperCAmelCase : list[list[int]] ) -> None: '''simple docstring''' for i in matrix: print(*_UpperCAmelCase ) if __name__ == "__main__": lowercase = make_matrix() print("""\norigin:\n""") print_matrix(matrix) print("""\nrotate 90 counterclockwise:\n""") print_matrix(rotate_aa(matrix)) lowercase = make_matrix() print("""\norigin:\n""") print_matrix(matrix) print("""\nrotate 180:\n""") print_matrix(rotate_aaa(matrix)) lowercase = make_matrix() print("""\norigin:\n""") print_matrix(matrix) print("""\nrotate 270 counterclockwise:\n""") print_matrix(rotate_aaa(matrix))
150
"""simple docstring""" import math import unittest def A__ ( _UpperCAmelCase : int ) -> bool: '''simple docstring''' assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True class SCREAMING_SNAKE_CASE_ ( unittest.TestCase): '''simple docstring''' def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.assertTrue(is_prime(2)) self.assertTrue(is_prime(3)) self.assertTrue(is_prime(5)) self.assertTrue(is_prime(7)) self.assertTrue(is_prime(11)) self.assertTrue(is_prime(13)) self.assertTrue(is_prime(17)) self.assertTrue(is_prime(19)) self.assertTrue(is_prime(23)) self.assertTrue(is_prime(29)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' with self.assertRaises(lowerCamelCase__): is_prime(-19) self.assertFalse( is_prime(0) , "Zero doesn't have any positive factors, primes must have exactly two." , ) self.assertFalse( is_prime(1) , "One only has 1 positive factor, primes must have exactly two." , ) self.assertFalse(is_prime(2 * 2)) self.assertFalse(is_prime(2 * 3)) self.assertFalse(is_prime(3 * 3)) self.assertFalse(is_prime(3 * 5)) self.assertFalse(is_prime(3 * 5 * 7)) if __name__ == "__main__": unittest.main()
150
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "google/pegasus-large": "https://huggingface.co/google/pegasus-large/resolve/main/config.json", # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class __a ( __lowerCamelCase ): """simple docstring""" _A : Optional[Any] = "pegasus" _A : List[str] = ["past_key_values"] _A : List[str] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Dict ,_UpperCamelCase : str=5_0_2_6_5 ,_UpperCamelCase : Any=1_0_2_4 ,_UpperCamelCase : List[Any]=1_2 ,_UpperCamelCase : List[Any]=4_0_9_6 ,_UpperCamelCase : Optional[Any]=1_6 ,_UpperCamelCase : Any=1_2 ,_UpperCamelCase : str=4_0_9_6 ,_UpperCamelCase : str=1_6 ,_UpperCamelCase : Optional[Any]=0.0 ,_UpperCamelCase : Tuple=0.0 ,_UpperCamelCase : Optional[Any]=True ,_UpperCamelCase : List[str]=True ,_UpperCamelCase : Optional[int]="gelu" ,_UpperCamelCase : Dict=1_0_2_4 ,_UpperCamelCase : str=0.1 ,_UpperCamelCase : List[str]=0.0 ,_UpperCamelCase : Dict=0.0 ,_UpperCamelCase : Tuple=0.02 ,_UpperCamelCase : Optional[Any]=0 ,_UpperCamelCase : Optional[int]=False ,_UpperCamelCase : Dict=0 ,_UpperCamelCase : List[str]=1 ,_UpperCamelCase : List[str]=1 ,**_UpperCamelCase : List[str] ,) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ =vocab_size SCREAMING_SNAKE_CASE__ =max_position_embeddings SCREAMING_SNAKE_CASE__ =d_model SCREAMING_SNAKE_CASE__ =encoder_ffn_dim SCREAMING_SNAKE_CASE__ =encoder_layers SCREAMING_SNAKE_CASE__ =encoder_attention_heads SCREAMING_SNAKE_CASE__ =decoder_ffn_dim SCREAMING_SNAKE_CASE__ =decoder_layers SCREAMING_SNAKE_CASE__ =decoder_attention_heads SCREAMING_SNAKE_CASE__ =dropout SCREAMING_SNAKE_CASE__ =attention_dropout SCREAMING_SNAKE_CASE__ =activation_dropout SCREAMING_SNAKE_CASE__ =activation_function SCREAMING_SNAKE_CASE__ =init_std SCREAMING_SNAKE_CASE__ =encoder_layerdrop SCREAMING_SNAKE_CASE__ =decoder_layerdrop SCREAMING_SNAKE_CASE__ =use_cache SCREAMING_SNAKE_CASE__ =encoder_layers SCREAMING_SNAKE_CASE__ =scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=_UpperCamelCase ,eos_token_id=_UpperCamelCase ,is_encoder_decoder=_UpperCamelCase ,decoder_start_token_id=_UpperCamelCase ,forced_eos_token_id=_UpperCamelCase ,**_UpperCamelCase ,) @property def __A ( self : Any ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def __A ( self : Any ) -> int: '''simple docstring''' return self.d_model
151
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase_ = { "configuration_rag": ["RagConfig"], "retrieval_rag": ["RagRetriever"], "tokenization_rag": ["RagTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ "RagModel", "RagPreTrainedModel", "RagSequenceForGeneration", "RagTokenForGeneration", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ "TFRagModel", "TFRagPreTrainedModel", "TFRagSequenceForGeneration", "TFRagTokenForGeneration", ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
151
1
"""simple docstring""" import random def A( snake_case_ , snake_case_ , snake_case_ = False ): """simple docstring""" lowercase__: dict = {i: [] for i in range(snake_case_ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(snake_case_ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(snake_case_ ): for j in range(i + 1 , snake_case_ ): if random.random() < probability: graph[i].append(snake_case_ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(snake_case_ ) return graph def A( snake_case_ ): """simple docstring""" return { i: [j for j in range(snake_case_ ) if i != j] for i in range(snake_case_ ) } if __name__ == "__main__": import doctest doctest.testmod()
120
"""simple docstring""" import os def A( ): """simple docstring""" with open(os.path.dirname(snake_case_ ) + "/p022_names.txt" ) as file: lowercase__: str = str(file.readlines()[0] ) lowercase__: str = names.replace("\"" , "" ).split("," ) names.sort() lowercase__: Any = 0 lowercase__: List[str] = 0 for i, name in enumerate(snake_case_ ): for letter in name: name_score += ord(snake_case_ ) - 64 total_score += (i + 1) * name_score lowercase__: Dict = 0 return total_score if __name__ == "__main__": print(solution())
120
1
'''simple docstring''' import os import re import shutil import sys import tempfile import unittest import black UpperCAmelCase_ : str = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. UpperCAmelCase_ : Dict = ' def __init__(self, config):\n super().__init__()\n self.transform = BertPredictionHeadTransform(config)\n\n # The output weights are the same as the input embeddings, but there is\n # an output-only bias for each token.\n self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)\n\n self.bias = nn.Parameter(torch.zeros(config.vocab_size))\n\n # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`\n self.decoder.bias = self.bias\n\n def forward(self, hidden_states):\n hidden_states = self.transform(hidden_states)\n hidden_states = self.decoder(hidden_states)\n return hidden_states\n' class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): """simple docstring""" __A : Union[str, Any] = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , 'models/bert/' ) ) __A : Optional[int] = self.transformer_dir shutil.copy( os.path.join(__lowercase , 'src/transformers/models/bert/modeling_bert.py' ) , os.path.join(self.transformer_dir , 'models/bert/modeling_bert.py' ) , ) def snake_case__ ( self ): """simple docstring""" __A : Any = 'src/transformers' shutil.rmtree(self.transformer_dir ) def snake_case__ ( self , __lowercase , __lowercase , __lowercase , __lowercase=None ): """simple docstring""" __A : Union[str, Any] = comment + F"""\nclass {class_name}(nn.Module):\n""" + class_code if overwrite_result is not None: __A : int = comment + F"""\nclass {class_name}(nn.Module):\n""" + overwrite_result __A : int = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) __A : int = black.format_str(__lowercase , mode=__lowercase ) __A : Tuple = os.path.join(self.transformer_dir , 'new_code.py' ) with open(__lowercase , 'w' , newline='\n' ) as f: f.write(__lowercase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(__lowercase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=__lowercase ) with open(__lowercase , 'r' ) as f: self.assertTrue(f.read() , __lowercase ) def snake_case__ ( self ): """simple docstring""" __A : int = check_copies.find_code_in_transformers('models.bert.modeling_bert.BertLMPredictionHead' ) self.assertEqual(__lowercase , __lowercase ) def snake_case__ ( self ): """simple docstring""" self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead' , 'BertLMPredictionHead' , REFERENCE_CODE + '\n' , ) # With no empty line at the end self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead' , 'BertLMPredictionHead' , __lowercase , ) # Copy consistency with rename self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel' , 'TestModelLMPredictionHead' , re.sub('Bert' , 'TestModel' , __lowercase ) , ) # Copy consistency with a really long name __A : Union[str, Any] = 'TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason' self.check_copy_consistency( F"""# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}""" , F"""{long_class_name}LMPredictionHead""" , re.sub('Bert' , __lowercase , __lowercase ) , ) # Copy consistency with overwrite self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel' , 'TestModelLMPredictionHead' , __lowercase , overwrite_result=re.sub('Bert' , 'TestModel' , __lowercase ) , ) def snake_case__ ( self ): """simple docstring""" __A : Optional[Any] = check_copies.LOCALIZED_READMES['README_zh-hans.md'] __A : Any = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the' ' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for' ' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong' ' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.' ' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),' ' released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and' ' lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same' ' method has been applied to compress GPT2 into' ' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into' ' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),' ' Multilingual BERT into' ' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German' ' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**' ' (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders' ' as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang' ' Luong, Quoc V. Le, Christopher D. Manning.' ) __A : List[str] = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) __A : Tuple = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.' ' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文' ' [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and' ' lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same' ' method has been applied to compress GPT2 into' ' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into' ' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),' ' Multilingual BERT into' ' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German' ' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自' ' Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather' ' than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,' ' Christopher D. Manning 发布。\n' ) __A ,__A : Union[str, Any] = check_copies.convert_to_localized_md( __lowercase , __lowercase , localized_readme['format_model_list'] ) self.assertFalse(__lowercase ) self.assertEqual(__lowercase , __lowercase ) __A ,__A : List[str] = check_copies.convert_to_localized_md( __lowercase , __lowercase , localized_readme['format_model_list'] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(__lowercase ) __A : Dict = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the' ' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for' ' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong' ' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.' ) __A : List[Any] = ( '1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and' ' the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) __A : int = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) __A ,__A : Any = check_copies.convert_to_localized_md( __lowercase , __lowercase , localized_readme['format_model_list'] ) # Check if the model link is synchronized. self.assertEqual(__lowercase , __lowercase )
365
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : int = logging.get_logger(__name__) UpperCAmelCase_ : List[Any] = { 'facebook/deit-base-distilled-patch16-224': ( 'https://huggingface.co/facebook/deit-base-patch16-224/resolve/main/config.json' ), # See all DeiT models at https://huggingface.co/models?filter=deit } class _lowerCamelCase ( snake_case_ ): '''simple docstring''' __lowercase : Union[str, Any] = '''deit''' def __init__( self , __lowercase=768 , __lowercase=12 , __lowercase=12 , __lowercase=3_072 , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.0_2 , __lowercase=1E-12 , __lowercase=224 , __lowercase=16 , __lowercase=3 , __lowercase=True , __lowercase=16 , **__lowercase , ): """simple docstring""" super().__init__(**__lowercase ) __A : Union[str, Any] = hidden_size __A : int = num_hidden_layers __A : str = num_attention_heads __A : Optional[int] = intermediate_size __A : List[str] = hidden_act __A : str = hidden_dropout_prob __A : str = attention_probs_dropout_prob __A : int = initializer_range __A : Tuple = layer_norm_eps __A : List[Any] = image_size __A : Dict = patch_size __A : Optional[Any] = num_channels __A : List[Any] = qkv_bias __A : str = encoder_stride class _lowerCamelCase ( snake_case_ ): '''simple docstring''' __lowercase : Optional[int] = version.parse('''1.11''' ) @property def snake_case__ ( self ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def snake_case__ ( self ): """simple docstring""" return 1E-4
365
1
"""simple docstring""" class _lowercase : """simple docstring""" def __init__( self : Tuple , UpperCamelCase__ : Any ) -> int: '''simple docstring''' __UpperCamelCase =arr.split(''',''' ) def UpperCAmelCase_ ( self : List[Any] ) -> int: '''simple docstring''' __UpperCamelCase =[int(self.array[0] )] * len(self.array ) __UpperCamelCase =[int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): __UpperCamelCase =max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) __UpperCamelCase =max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": __lowercase = input('''please input some numbers:''') __lowercase = SubArray(whole_array) __lowercase = array.solve_sub_array() print(('''the results is:''', re))
296
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class _lowercase ( ctypes.Structure ): """simple docstring""" lowercase__ = [('''size''', ctypes.c_int), ('''visible''', ctypes.c_byte)] def lowerCAmelCase (): """simple docstring""" if os.name == "nt": __UpperCamelCase =CursorInfo() __UpperCamelCase =ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__UpperCamelCase , ctypes.byref(__UpperCamelCase ) ) __UpperCamelCase =False ctypes.windll.kernelaa.SetConsoleCursorInfo(__UpperCamelCase , ctypes.byref(__UpperCamelCase ) ) elif os.name == "posix": sys.stdout.write('''\033[?25l''' ) sys.stdout.flush() def lowerCAmelCase (): """simple docstring""" if os.name == "nt": __UpperCamelCase =CursorInfo() __UpperCamelCase =ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__UpperCamelCase , ctypes.byref(__UpperCamelCase ) ) __UpperCamelCase =True ctypes.windll.kernelaa.SetConsoleCursorInfo(__UpperCamelCase , ctypes.byref(__UpperCamelCase ) ) elif os.name == "posix": sys.stdout.write('''\033[?25h''' ) sys.stdout.flush() @contextmanager def lowerCAmelCase (): """simple docstring""" try: hide_cursor() yield finally: show_cursor()
296
1
def UpperCamelCase ( snake_case__ = 1 , snake_case__ = 10_00): lowerCAmelCase_ : str = 1 lowerCAmelCase_ : Union[str, Any] = 0 for divide_by_number in range(snake_case__ , digit + 1): lowerCAmelCase_ : list[int] = [] lowerCAmelCase_ : Union[str, Any] = numerator for _ in range(1 , digit + 1): if now_divide in has_been_divided: if longest_list_length < len(snake_case__): lowerCAmelCase_ : List[str] = len(snake_case__) lowerCAmelCase_ : Any = divide_by_number else: has_been_divided.append(snake_case__) lowerCAmelCase_ : List[str] = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
659
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCamelCase ( ): lowerCAmelCase_ : Dict = HfArgumentParser(snake_case__) lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses()[0] lowerCAmelCase_ : List[Any] = TensorFlowBenchmark(args=snake_case__) try: lowerCAmelCase_ : str = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowerCAmelCase_ : Optional[Any] = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowerCAmelCase_ : Tuple = " ".join(str(snake_case__).split(" ")[:-1]) lowerCAmelCase_ : List[Any] = "" lowerCAmelCase_ : Optional[Any] = eval(str(snake_case__).split(" ")[-1]) lowerCAmelCase_ : List[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:]) else: wrong_args.append(snake_case__) if len(snake_case__) > 0: lowerCAmelCase_ : int = full_error_msg + begin_error_msg + str(snake_case__) raise ValueError(snake_case__) benchmark.run() if __name__ == "__main__": main()
659
1
from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class UpperCAmelCase__ ( a__ ): """simple docstring""" UpperCAmelCase__ : Tuple = DistilBertTokenizer UpperCAmelCase__ : Any = DistilBertTokenizerFast UpperCAmelCase__ : Dict = True @slow def _a ( self ) -> Union[str, Any]: __UpperCamelCase =DistilBertTokenizer.from_pretrained('distilbert-base-uncased' ) __UpperCamelCase =tokenizer.encode('sequence builders' , add_special_tokens=lowerCamelCase_ ) __UpperCamelCase =tokenizer.encode('multi-sequence build' , add_special_tokens=lowerCamelCase_ ) __UpperCamelCase =tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ ) __UpperCamelCase =tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ , lowerCamelCase_ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
710
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self , A_ , A_=7 , A_=3 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , ) -> List[Any]: __UpperCamelCase =size if size is not None else {'height': 18, 'width': 18} __UpperCamelCase =parent __UpperCamelCase =batch_size __UpperCamelCase =num_channels __UpperCamelCase =image_size __UpperCamelCase =min_resolution __UpperCamelCase =max_resolution __UpperCamelCase =do_resize __UpperCamelCase =size __UpperCamelCase =apply_ocr def _a ( self ) -> Tuple: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class UpperCAmelCase__ ( A_ , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : List[str] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def _a ( self ) -> Optional[Any]: __UpperCamelCase =LayoutLMvaImageProcessingTester(self ) @property def _a ( self ) -> Union[str, Any]: return self.image_processor_tester.prepare_image_processor_dict() def _a ( self ) -> List[Any]: __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) self.assertTrue(hasattr(A_ , 'apply_ocr' ) ) def _a ( self ) -> Dict: __UpperCamelCase =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 18, 'width': 18} ) __UpperCamelCase =self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'height': 42, 'width': 42} ) def _a ( self ) -> Dict: pass def _a ( self ) -> Optional[Any]: # Initialize image_processing __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) # create random PIL images __UpperCamelCase =prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input __UpperCamelCase =image_processing(image_inputs[0] , return_tensors='pt' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) self.assertIsInstance(encoding.words , A_ ) self.assertIsInstance(encoding.boxes , A_ ) # Test batched __UpperCamelCase =image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def _a ( self ) -> int: # Initialize image_processing __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __UpperCamelCase =prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input __UpperCamelCase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) # Test batched __UpperCamelCase =image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def _a ( self ) -> List[str]: # Initialize image_processing __UpperCamelCase =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __UpperCamelCase =prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input __UpperCamelCase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) # Test batched __UpperCamelCase =image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def _a ( self ) -> Any: # with apply_OCR = True __UpperCamelCase =LayoutLMvaImageProcessor() from datasets import load_dataset __UpperCamelCase =load_dataset('hf-internal-testing/fixtures_docvqa' , split='test' ) __UpperCamelCase =Image.open(ds[0]['file'] ).convert('RGB' ) __UpperCamelCase =image_processing(A_ , return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __UpperCamelCase =[['11:14', 'to', '11:39', 'a.m', '11:39', 'to', '11:44', 'a.m.', '11:44', 'a.m.', 'to', '12:25', 'p.m.', '12:25', 'to', '12:58', 'p.m.', '12:58', 'to', '4:00', 'p.m.', '2:00', 'to', '5:00', 'p.m.', 'Coffee', 'Break', 'Coffee', 'will', 'be', 'served', 'for', 'men', 'and', 'women', 'in', 'the', 'lobby', 'adjacent', 'to', 'exhibit', 'area.', 'Please', 'move', 'into', 'exhibit', 'area.', '(Exhibits', 'Open)', 'TRRF', 'GENERAL', 'SESSION', '(PART', '|)', 'Presiding:', 'Lee', 'A.', 'Waller', 'TRRF', 'Vice', 'President', '“Introductory', 'Remarks”', 'Lee', 'A.', 'Waller,', 'TRRF', 'Vice', 'Presi-', 'dent', 'Individual', 'Interviews', 'with', 'TRRF', 'Public', 'Board', 'Members', 'and', 'Sci-', 'entific', 'Advisory', 'Council', 'Mem-', 'bers', 'Conducted', 'by', 'TRRF', 'Treasurer', 'Philip', 'G.', 'Kuehn', 'to', 'get', 'answers', 'which', 'the', 'public', 'refrigerated', 'warehousing', 'industry', 'is', 'looking', 'for.', 'Plus', 'questions', 'from', 'the', 'floor.', 'Dr.', 'Emil', 'M.', 'Mrak,', 'University', 'of', 'Cal-', 'ifornia,', 'Chairman,', 'TRRF', 'Board;', 'Sam', 'R.', 'Cecil,', 'University', 'of', 'Georgia', 'College', 'of', 'Agriculture;', 'Dr.', 'Stanley', 'Charm,', 'Tufts', 'University', 'School', 'of', 'Medicine;', 'Dr.', 'Robert', 'H.', 'Cotton,', 'ITT', 'Continental', 'Baking', 'Company;', 'Dr.', 'Owen', 'Fennema,', 'University', 'of', 'Wis-', 'consin;', 'Dr.', 'Robert', 'E.', 'Hardenburg,', 'USDA.', 'Questions', 'and', 'Answers', 'Exhibits', 'Open', 'Capt.', 'Jack', 'Stoney', 'Room', 'TRRF', 'Scientific', 'Advisory', 'Council', 'Meeting', 'Ballroom', 'Foyer']] # noqa: E231 __UpperCamelCase =[[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , A_ ) self.assertListEqual(encoding.boxes , A_ ) # with apply_OCR = False __UpperCamelCase =LayoutLMvaImageProcessor(apply_ocr=A_ ) __UpperCamelCase =image_processing(A_ , return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
682
0
def __lowerCAmelCase( _SCREAMING_SNAKE_CASE ) -> bool: """simple docstring""" if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): raise ValueError('check_bouncy() accepts only integer arguments' ) _A = str(_SCREAMING_SNAKE_CASE ) _A = ''.join(sorted(_SCREAMING_SNAKE_CASE ) ) return sorted_str_n != str_n and sorted_str_n[::-1] != str_n def __lowerCAmelCase( _SCREAMING_SNAKE_CASE = 99 ) -> int: """simple docstring""" if not 0 < percent < 100: raise ValueError('solution() only accepts values from 0 to 100' ) _A = 0 _A = 1 while True: if check_bouncy(_SCREAMING_SNAKE_CASE ): bouncy_num += 1 if (bouncy_num / num) * 100 >= percent: return num num += 1 if __name__ == "__main__": from doctest import testmod testmod() print(f"{solution(99)}")
27
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase = { 'configuration_mobilebert': [ 'MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileBertConfig', 'MobileBertOnnxConfig', ], 'tokenization_mobilebert': ['MobileBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ['MobileBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ 'MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileBertForMaskedLM', 'MobileBertForMultipleChoice', 'MobileBertForNextSentencePrediction', 'MobileBertForPreTraining', 'MobileBertForQuestionAnswering', 'MobileBertForSequenceClassification', 'MobileBertForTokenClassification', 'MobileBertLayer', 'MobileBertModel', 'MobileBertPreTrainedModel', 'load_tf_weights_in_mobilebert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ 'TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFMobileBertForMaskedLM', 'TFMobileBertForMultipleChoice', 'TFMobileBertForNextSentencePrediction', 'TFMobileBertForPreTraining', 'TFMobileBertForQuestionAnswering', 'TFMobileBertForSequenceClassification', 'TFMobileBertForTokenClassification', 'TFMobileBertMainLayer', 'TFMobileBertModel', 'TFMobileBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
264
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase : Tuple = logging.get_logger(__name__) __UpperCAmelCase : Optional[Any] = {} class _snake_case ( _A ): _A = 'llama' _A = ['past_key_values'] def __init__( self ,UpperCamelCase=32_000 ,UpperCamelCase=4_096 ,UpperCamelCase=11_008 ,UpperCamelCase=32 ,UpperCamelCase=32 ,UpperCamelCase=None ,UpperCamelCase="silu" ,UpperCamelCase=2_048 ,UpperCamelCase=0.02 ,UpperCamelCase=1E-6 ,UpperCamelCase=True ,UpperCamelCase=0 ,UpperCamelCase=1 ,UpperCamelCase=2 ,UpperCamelCase=1 ,UpperCamelCase=False ,UpperCamelCase=None ,**UpperCamelCase ,) -> Tuple: snake_case__ :int = vocab_size snake_case__ :Any = max_position_embeddings snake_case__ :int = hidden_size snake_case__ :List[Any] = intermediate_size snake_case__ :int = num_hidden_layers snake_case__ :Union[str, Any] = num_attention_heads # for backward compatibility if num_key_value_heads is None: snake_case__ :Dict = num_attention_heads snake_case__ :List[str] = num_key_value_heads snake_case__ :Optional[int] = hidden_act snake_case__ :Any = initializer_range snake_case__ :Dict = rms_norm_eps snake_case__ :List[str] = pretraining_tp snake_case__ :Any = use_cache snake_case__ :Optional[Any] = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=UpperCamelCase ,bos_token_id=UpperCamelCase ,eos_token_id=UpperCamelCase ,tie_word_embeddings=UpperCamelCase ,**UpperCamelCase ,) def lowerCAmelCase_ ( self ) -> Optional[int]: if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,UpperCamelCase ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " f'got {self.rope_scaling}' ) snake_case__ :Optional[int] = self.rope_scaling.get("type" ,UpperCamelCase ) snake_case__ :Optional[Any] = self.rope_scaling.get("factor" ,UpperCamelCase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}' ) if rope_scaling_factor is None or not isinstance(UpperCamelCase ,UpperCamelCase ) or rope_scaling_factor <= 1.0: raise ValueError(f'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}' )
719
def lowercase_ ( __snake_case : list ) -> list: '''simple docstring''' if any(not isinstance(__snake_case , __snake_case ) or x < 0 for x in sequence ): raise TypeError("Sequence must be list of non-negative integers" ) for _ in range(len(__snake_case ) ): for i, (rod_upper, rod_lower) in enumerate(zip(__snake_case , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
57
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''EleutherAI/gpt-neox-20b''': '''https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json''', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class A ( __UpperCAmelCase ): __snake_case = 'gpt_neox' def __init__( self, UpperCamelCase__=5_0432, UpperCamelCase__=6144, UpperCamelCase__=44, UpperCamelCase__=64, UpperCamelCase__=2_4576, UpperCamelCase__="gelu", UpperCamelCase__=0.25, UpperCamelCase__=1_0000, UpperCamelCase__=0.0, UpperCamelCase__=0.0, UpperCamelCase__=0.1, UpperCamelCase__=2048, UpperCamelCase__=0.02, UpperCamelCase__=1E-5, UpperCamelCase__=True, UpperCamelCase__=0, UpperCamelCase__=2, UpperCamelCase__=False, UpperCamelCase__=True, UpperCamelCase__=None, **UpperCamelCase__, ): """simple docstring""" super().__init__(bos_token_id=UpperCamelCase__, eos_token_id=UpperCamelCase__, **UpperCamelCase__ ) lowerCAmelCase_ = vocab_size lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_act lowerCAmelCase_ = rotary_pct lowerCAmelCase_ = rotary_emb_base lowerCAmelCase_ = attention_dropout lowerCAmelCase_ = hidden_dropout lowerCAmelCase_ = classifier_dropout lowerCAmelCase_ = initializer_range lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = use_cache lowerCAmelCase_ = tie_word_embeddings lowerCAmelCase_ = use_parallel_residual lowerCAmelCase_ = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( '''The hidden size is not divisble by the number of attention heads! Make sure to update them!''' ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling, UpperCamelCase__ ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' f"got {self.rope_scaling}" ) lowerCAmelCase_ = self.rope_scaling.get('''type''', UpperCamelCase__ ) lowerCAmelCase_ = self.rope_scaling.get('''factor''', UpperCamelCase__ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(UpperCamelCase__, UpperCamelCase__ ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
431
_A = '''Alexander Joslin''' import operator as op from .stack import Stack def __UpperCamelCase ( _A ): lowerCAmelCase_ = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} lowerCAmelCase_ = Stack() lowerCAmelCase_ = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(_A ) ) elif i in operators: # RULE 2 operator_stack.push(_A ) elif i == ")": # RULE 4 lowerCAmelCase_ = operator_stack.peek() operator_stack.pop() lowerCAmelCase_ = operand_stack.peek() operand_stack.pop() lowerCAmelCase_ = operand_stack.peek() operand_stack.pop() lowerCAmelCase_ = operators[opr](_A , _A ) operand_stack.push(_A ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": _A = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(f"{equation} = {dijkstras_two_stack_algorithm(equation)}")
431
1
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' _SCREAMING_SNAKE_CASE : List[Any] = IFImgaImgSuperResolutionPipeline _SCREAMING_SNAKE_CASE : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'width', 'height'} _SCREAMING_SNAKE_CASE : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"original_image"} ) _SCREAMING_SNAKE_CASE : Optional[Any] = PipelineTesterMixin.required_optional_params - {'latents'} def snake_case__ ( self) -> Optional[int]: """simple docstring""" return self._get_superresolution_dummy_components() def snake_case__ ( self , _A , _A=0) -> Optional[Any]: """simple docstring""" if str(A_).startswith('''mps'''): _UpperCAmelCase : List[str] = torch.manual_seed(A_) else: _UpperCAmelCase : Optional[int] = torch.Generator(device=A_).manual_seed(A_) _UpperCAmelCase : Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(A_)).to(A_) _UpperCAmelCase : Tuple = floats_tensor((1, 3, 16, 16) , rng=random.Random(A_)).to(A_) _UpperCAmelCase : Any = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def snake_case__ ( self) -> List[str]: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3) def snake_case__ ( self) -> List[str]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''') def snake_case__ ( self) -> int: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1e-1) def snake_case__ ( self) -> List[Any]: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1e-2) def snake_case__ ( self) -> Optional[Any]: """simple docstring""" self._test_save_load_local() def snake_case__ ( self) -> List[Any]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
716
import os from collections.abc import Iterator def _lowerCamelCase ( __A : str = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(__A ): _UpperCAmelCase : List[Any] = [d for d in dir_names if d != '''scripts''' and d[0] not in '''._'''] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(__A )[1] in (".py", ".ipynb"): yield os.path.join(__A , __A ).lstrip('''./''' ) def _lowerCamelCase ( __A : Dict ) -> List[Any]: return f'''{i * ' '}*''' if i else "\n##" def _lowerCamelCase ( __A : str , __A : str ) -> str: _UpperCAmelCase : int = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(__A ) or old_parts[i] != new_part) and new_part: print(f'''{md_prefix(__A )} {new_part.replace('_' , ' ' ).title()}''' ) return new_path def _lowerCamelCase ( __A : str = "." ) -> None: _UpperCAmelCase : List[str] = '''''' for filepath in sorted(good_file_paths(__A ) ): _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = os.path.split(__A ) if filepath != old_path: _UpperCAmelCase : Optional[int] = print_path(__A , __A ) _UpperCAmelCase : Dict = (filepath.count(os.sep ) + 1) if filepath else 0 _UpperCAmelCase : Union[str, Any] = f'''{filepath}/{filename}'''.replace(''' ''' , '''%20''' ) _UpperCAmelCase : List[Any] = os.path.splitext(filename.replace('''_''' , ''' ''' ).title() )[0] print(f'''{md_prefix(__A )} [{filename}]({url})''' ) if __name__ == "__main__": print_directory_md('.')
186
0
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_ ( UpperCamelCase__ : str = "laptop" ): """simple docstring""" __lowercase = f'''https://www.amazon.in/laptop/s?k={product}''' __lowercase = { """User-Agent""": """Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36""", """Accept-Language""": """en-US, en;q=0.5""", } __lowercase = BeautifulSoup(requests.get(UpperCamelCase__ , headers=UpperCamelCase__ ).text ) # Initialize a Pandas dataframe with the column titles __lowercase = DataFrame( columns=[ """Product Title""", """Product Link""", """Current Price of the product""", """Product Rating""", """MRP of the product""", """Discount""", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( """div""" , attrs={"""class""": """s-result-item""", """data-component-type""": """s-search-result"""} , ) , soup.find_all("""div""" , attrs={"""class""": """a-row a-size-base a-color-base"""} ) , ): try: __lowercase = item.ha.text __lowercase = """https://www.amazon.in/""" + item.ha.a["""href"""] __lowercase = item.find("""span""" , attrs={"""class""": """a-offscreen"""} ).text try: __lowercase = item.find("""span""" , attrs={"""class""": """a-icon-alt"""} ).text except AttributeError: __lowercase = """Not available""" try: __lowercase = ( """₹""" + item.find( """span""" , attrs={"""class""": """a-price a-text-price"""} ).text.split("""₹""" )[1] ) except AttributeError: __lowercase = """""" try: __lowercase = float( ( ( float(product_mrp.strip("""₹""" ).replace(""",""" , """""" ) ) - float(product_price.strip("""₹""" ).replace(""",""" , """""" ) ) ) / float(product_mrp.strip("""₹""" ).replace(""",""" , """""" ) ) ) * 100 ) except ValueError: __lowercase = float("""nan""" ) except AttributeError: pass __lowercase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __lowercase = """ """ __lowercase = """ """ data_frame.index += 1 return data_frame if __name__ == "__main__": UpperCAmelCase__ ="headphones" get_amazon_product_data(product).to_csv(f"""Amazon Product Data for {product}.csv""")
616
"""simple docstring""" def lowerCAmelCase_ ( UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" return x if y == 0 else greatest_common_divisor(UpperCamelCase__ , x % y ) def lowerCAmelCase_ ( UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" return (x * y) // greatest_common_divisor(UpperCamelCase__ , UpperCamelCase__ ) def lowerCAmelCase_ ( UpperCamelCase__ : int = 20 ): """simple docstring""" __lowercase = 1 for i in range(1 , n + 1 ): __lowercase = lcm(UpperCamelCase__ , UpperCamelCase__ ) return g if __name__ == "__main__": print(f"""{solution() = }""")
616
1
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase__ : '''simple docstring''' @staticmethod def snake_case__ ( *a_ : List[str] , **a_ : str ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_torch class UpperCAmelCase__ ( unittest.TestCase ): '''simple docstring''' UpperCamelCase = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def snake_case__ ( self : Any , a_ : Dict , a_ : int , a_ : int ): '''simple docstring''' __UpperCAmelCase : Dict = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) __UpperCAmelCase : Any = [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] return object_detector, examples def snake_case__ ( self : List[Any] , a_ : Union[str, Any] , a_ : Optional[int] ): '''simple docstring''' __UpperCAmelCase : Tuple = object_detector(examples[0] , threshold=0.0 ) __UpperCAmelCase : Optional[Any] = len(a_ ) self.assertGreater(a_ , 0 ) self.assertEqual( a_ , [ { '''score''': ANY(a_ ), '''label''': ANY(a_ ), '''box''': {'''xmin''': ANY(a_ ), '''ymin''': ANY(a_ ), '''xmax''': ANY(a_ ), '''ymax''': ANY(a_ )}, } for i in range(a_ ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def snake_case__ ( self : Tuple ): '''simple docstring''' pass @require_torch def snake_case__ ( self : Any ): '''simple docstring''' __UpperCAmelCase : List[Any] = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) __UpperCAmelCase : List[Any] = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.6_4 , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {'''score''': 0.7_2_3_5, '''label''': '''cat''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_2_1_8, '''label''': '''remote''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_1_8_4, '''label''': '''couch''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.6_7_4_8, '''label''': '''remote''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_5_6, '''label''': '''cat''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_1_4, '''label''': '''couch''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_4_5_6, '''label''': '''remote''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, {'''score''': 0.6_4_2, '''label''': '''remote''', '''box''': {'''xmin''': 67, '''ymin''': 2_74, '''xmax''': 93, '''ymax''': 2_97}}, {'''score''': 0.6_4_1_9, '''label''': '''cat''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, ] , ) __UpperCAmelCase : int = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.6_4 , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ [ {'''score''': 0.7_2_3_5, '''label''': '''cat''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_2_1_8, '''label''': '''remote''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_1_8_4, '''label''': '''couch''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.6_7_4_8, '''label''': '''remote''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_5_6, '''label''': '''cat''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_1_4, '''label''': '''couch''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_4_5_6, '''label''': '''remote''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, {'''score''': 0.6_4_2, '''label''': '''remote''', '''box''': {'''xmin''': 67, '''ymin''': 2_74, '''xmax''': 93, '''ymax''': 2_97}}, {'''score''': 0.6_4_1_9, '''label''': '''cat''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, ] ] , ) @require_torch @slow def snake_case__ ( self : Tuple ): '''simple docstring''' __UpperCAmelCase : Any = pipeline('''zero-shot-object-detection''' ) __UpperCAmelCase : Dict = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, {'''score''': 0.1_4_7_4, '''label''': '''remote''', '''box''': {'''xmin''': 3_35, '''ymin''': 74, '''xmax''': 3_71, '''ymax''': 1_87}}, {'''score''': 0.1_2_0_8, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_42, '''ymax''': 4_76}}, ] , ) __UpperCAmelCase : int = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, {'''score''': 0.1_4_7_4, '''label''': '''remote''', '''box''': {'''xmin''': 3_35, '''ymin''': 74, '''xmax''': 3_71, '''ymax''': 1_87}}, {'''score''': 0.1_2_0_8, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_42, '''ymax''': 4_76}}, ], [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, {'''score''': 0.1_4_7_4, '''label''': '''remote''', '''box''': {'''xmin''': 3_35, '''ymin''': 74, '''xmax''': 3_71, '''ymax''': 1_87}}, {'''score''': 0.1_2_0_8, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_42, '''ymax''': 4_76}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def snake_case__ ( self : List[Any] ): '''simple docstring''' pass @require_torch @slow def snake_case__ ( self : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = 0.2 __UpperCAmelCase : List[Any] = pipeline('''zero-shot-object-detection''' ) __UpperCAmelCase : Any = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=a_ , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, ] , ) @require_torch @slow def snake_case__ ( self : int ): '''simple docstring''' __UpperCAmelCase : Tuple = 2 __UpperCAmelCase : Optional[int] = pipeline('''zero-shot-object-detection''' ) __UpperCAmelCase : Any = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=a_ , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, ] , )
703
import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": __A ="%20".join(argv[1:]) if len(argv) > 1 else quote(str(input("Search: "))) print("Googling.....") __A =f'''https://www.google.com/search?q={query}&num=100''' __A =requests.get( url, headers={"User-Agent": str(UserAgent().random)}, ) try: __A =( BeautifulSoup(res.text, "html.parser") .find("div", attrs={"class": "yuRUbf"}) .find("a") .get("href") ) except AttributeError: __A =parse_qs( BeautifulSoup(res.text, "html.parser") .find("div", attrs={"class": "kCrYT"}) .find("a") .get("href") )["url"][0] webbrowser.open(link)
241
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __snake_case :Dict ={'configuration_van': ['VAN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'VanConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case :List[Any] =[ 'VAN_PRETRAINED_MODEL_ARCHIVE_LIST', 'VanForImageClassification', 'VanModel', 'VanPreTrainedModel', ] if TYPE_CHECKING: from .configuration_van import VAN_PRETRAINED_CONFIG_ARCHIVE_MAP, VanConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_van import ( VAN_PRETRAINED_MODEL_ARCHIVE_LIST, VanForImageClassification, VanModel, VanPreTrainedModel, ) else: import sys __snake_case :str =_LazyModule(__name__, globals()['__file__'], _import_structure)
106
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : list[int] ) -> list[int]: UpperCAmelCase_ = len(__UpperCamelCase ) for i in range(__UpperCamelCase ): for j in range(i + 1 , __UpperCamelCase ): if numbers[j] < numbers[i]: UpperCAmelCase_ , UpperCAmelCase_ = numbers[j], numbers[i] return numbers if __name__ == "__main__": _lowerCamelCase = input('Enter numbers separated by a comma:\n').strip() _lowerCamelCase = [int(item) for item in user_input.split(',')] print(exchange_sort(unsorted))
144
0
'''simple docstring''' lowercase_ = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on lowercase_ = {value: key for key, value in MORSE_CODE_DICT.items()} def UpperCamelCase__ ( a__ ): '''simple docstring''' return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def UpperCamelCase__ ( a__ ): '''simple docstring''' return "".join(REVERSE_DICT[char] for char in message.split() ) def UpperCamelCase__ ( ): '''simple docstring''' _lowerCAmelCase ='Morse code here!' print(a__ ) _lowerCAmelCase =encrypt(a__ ) print(a__ ) _lowerCAmelCase =decrypt(a__ ) print(a__ ) if __name__ == "__main__": main()
58
'''simple docstring''' import unittest from knapsack import knapsack as k class SCREAMING_SNAKE_CASE ( unittest.TestCase): """simple docstring""" def UpperCamelCase__ ( self ) -> Optional[Any]: _lowerCAmelCase =0 _lowerCAmelCase =[0] _lowerCAmelCase =[0] _lowerCAmelCase =len(__A ) self.assertEqual(k.knapsack(__A , __A , __A , __A ) , 0 ) _lowerCAmelCase =[60] _lowerCAmelCase =[10] _lowerCAmelCase =len(__A ) self.assertEqual(k.knapsack(__A , __A , __A , __A ) , 0 ) def UpperCamelCase__ ( self ) -> Tuple: _lowerCAmelCase =3 _lowerCAmelCase =[1, 2, 3] _lowerCAmelCase =[3, 2, 1] _lowerCAmelCase =len(__A ) self.assertEqual(k.knapsack(__A , __A , __A , __A ) , 5 ) def UpperCamelCase__ ( self ) -> Union[str, Any]: _lowerCAmelCase =50 _lowerCAmelCase =[60, 100, 120] _lowerCAmelCase =[10, 20, 30] _lowerCAmelCase =len(__A ) self.assertEqual(k.knapsack(__A , __A , __A , __A ) , 220 ) if __name__ == "__main__": unittest.main()
58
1
# using dfs for finding eulerian path traversal def __lowerCAmelCase ( _A ,_A ,_A ,_A=None ): """simple docstring""" _lowercase = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: _lowercase , _lowercase = True, True _lowercase = dfs(_A ,_A ,_A ,_A ) return path def __lowerCAmelCase ( _A ,_A ): """simple docstring""" _lowercase = 0 _lowercase = -1 for i in range(_A ): if i not in graph.keys(): continue if len(graph[i] ) % 2 == 1: odd_degree_nodes += 1 _lowercase = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def __lowerCAmelCase ( _A ,_A ): """simple docstring""" _lowercase = [[False for _ in range(max_node + 1 )] for _ in range(max_node + 1 )] _lowercase , _lowercase = check_circuit_or_path(_A ,_A ) if check == 3: print("""graph is not Eulerian""" ) print("""no path""" ) return _lowercase = 1 if check == 2: _lowercase = odd_node print("""graph has a Euler path""" ) if check == 1: print("""graph has a Euler cycle""" ) _lowercase = dfs(_A ,_A ,_A ) print(_A ) def __lowerCAmelCase ( ): """simple docstring""" _lowercase = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} _lowercase = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} _lowercase = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} _lowercase = {1: [2, 3], 2: [1, 3], 3: [1, 2]} _lowercase = { 1: [], 2: [] # all degree is zero } _lowercase = 10 check_euler(_A ,_A ) check_euler(_A ,_A ) check_euler(_A ,_A ) check_euler(_A ,_A ) check_euler(_A ,_A ) if __name__ == "__main__": main()
398
class _lowercase : """simple docstring""" def __init__( self , UpperCAmelCase ): '''simple docstring''' _lowercase = arr.split(""",""" ) def _UpperCAmelCase ( self ): '''simple docstring''' _lowercase = [int(self.array[0] )] * len(self.array ) _lowercase = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): _lowercase = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) _lowercase = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": A_: List[str] = input('please input some numbers:') A_: Optional[int] = SubArray(whole_array) A_: Dict = array.solve_sub_array() print(('the results is:', re))
398
1
"""simple docstring""" import os def lowerCamelCase__ ( ): '''simple docstring''' with open(os.path.dirname(_lowerCamelCase ) + '/p022_names.txt' ) as file: _lowerCAmelCase : List[str] = str(file.readlines()[0] ) _lowerCAmelCase : Optional[Any] = names.replace('"' , '' ).split(',' ) names.sort() _lowerCAmelCase : Dict = 0 _lowerCAmelCase : Union[str, Any] = 0 for i, name in enumerate(_lowerCamelCase ): for letter in name: name_score += ord(_lowerCamelCase ) - 64 total_score += (i + 1) * name_score _lowerCAmelCase : Dict = 0 return total_score if __name__ == "__main__": print(solution())
16
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class __UpperCamelCase ( a__ ): _UpperCAmelCase = 42 class __UpperCamelCase ( a__ , a__ ): @register_to_config def __init__( self ,_A = 32 ,_A = 64 ,_A = 20 ,_A = 768 ,_A=77 ,_A=4 ,_A = 0.0 ,_A = "silu" ,_A = None ,_A = None ,_A = "linear" ,_A = "prd" ,_A = None ,_A = None ,_A = None ,): '''simple docstring''' super().__init__() _lowerCAmelCase : Any = num_attention_heads _lowerCAmelCase : Optional[int] = attention_head_dim _lowerCAmelCase : Tuple = num_attention_heads * attention_head_dim _lowerCAmelCase : Optional[Any] = additional_embeddings _lowerCAmelCase : Union[str, Any] = time_embed_dim or inner_dim _lowerCAmelCase : Union[str, Any] = embedding_proj_dim or embedding_dim _lowerCAmelCase : Optional[int] = clip_embed_dim or embedding_dim _lowerCAmelCase : int = Timesteps(_A ,_A ,0 ) _lowerCAmelCase : int = TimestepEmbedding(_A ,_A ,out_dim=_A ,act_fn=_A ) _lowerCAmelCase : List[Any] = nn.Linear(_A ,_A ) if embedding_proj_norm_type is None: _lowerCAmelCase : Optional[Any] = None elif embedding_proj_norm_type == "layer": _lowerCAmelCase : List[Any] = nn.LayerNorm(_A ) else: raise ValueError(F"""unsupported embedding_proj_norm_type: {embedding_proj_norm_type}""" ) _lowerCAmelCase : Tuple = nn.Linear(_A ,_A ) if encoder_hid_proj_type is None: _lowerCAmelCase : int = None elif encoder_hid_proj_type == "linear": _lowerCAmelCase : List[Any] = nn.Linear(_A ,_A ) else: raise ValueError(F"""unsupported encoder_hid_proj_type: {encoder_hid_proj_type}""" ) _lowerCAmelCase : Dict = nn.Parameter(torch.zeros(1 ,num_embeddings + additional_embeddings ,_A ) ) if added_emb_type == "prd": _lowerCAmelCase : Dict = nn.Parameter(torch.zeros(1 ,1 ,_A ) ) elif added_emb_type is None: _lowerCAmelCase : List[Any] = None else: raise ValueError( F"""`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `'prd'` or `None`.""" ) _lowerCAmelCase : List[Any] = nn.ModuleList( [ BasicTransformerBlock( _A ,_A ,_A ,dropout=_A ,activation_fn='gelu' ,attention_bias=_A ,) for d in range(_A ) ] ) if norm_in_type == "layer": _lowerCAmelCase : Any = nn.LayerNorm(_A ) elif norm_in_type is None: _lowerCAmelCase : Any = None else: raise ValueError(F"""Unsupported norm_in_type: {norm_in_type}.""" ) _lowerCAmelCase : Union[str, Any] = nn.LayerNorm(_A ) _lowerCAmelCase : int = nn.Linear(_A ,_A ) _lowerCAmelCase : Any = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] ,-1_0_0_0_0.0 ) causal_attention_mask.triu_(1 ) _lowerCAmelCase : Tuple = causal_attention_mask[None, ...] self.register_buffer('causal_attention_mask' ,_A ,persistent=_A ) _lowerCAmelCase : Tuple = nn.Parameter(torch.zeros(1 ,_A ) ) _lowerCAmelCase : Dict = nn.Parameter(torch.zeros(1 ,_A ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def __lowerCamelCase ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = {} def fn_recursive_add_processors(_A ,_A ,_A ): if hasattr(_A ,'set_processor' ): _lowerCAmelCase : str = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"""{name}.{sub_name}""" ,_A ,_A ) return processors for name, module in self.named_children(): fn_recursive_add_processors(_A ,_A ,_A ) return processors def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : Optional[int] = len(self.attn_processors.keys() ) if isinstance(_A ,_A ) and len(_A ) != count: raise ValueError( F"""A dict of processors was passed, but the number of processors {len(_A )} does not match the""" F""" number of attention layers: {count}. Please make sure to pass {count} processor classes.""" ) def fn_recursive_attn_processor(_A ,_A ,_A ): if hasattr(_A ,'set_processor' ): if not isinstance(_A ,_A ): module.set_processor(_A ) else: module.set_processor(processor.pop(F"""{name}.processor""" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"""{name}.{sub_name}""" ,_A ,_A ) for name, module in self.named_children(): fn_recursive_attn_processor(_A ,_A ,_A ) def __lowerCamelCase ( self ): '''simple docstring''' self.set_attn_processor(AttnProcessor() ) def __lowerCamelCase ( self ,_A ,_A ,_A ,_A = None ,_A = None ,_A = True ,): '''simple docstring''' _lowerCAmelCase : str = hidden_states.shape[0] _lowerCAmelCase : int = timestep if not torch.is_tensor(_A ): _lowerCAmelCase : str = torch.tensor([timesteps] ,dtype=torch.long ,device=hidden_states.device ) elif torch.is_tensor(_A ) and len(timesteps.shape ) == 0: _lowerCAmelCase : Dict = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML _lowerCAmelCase : Optional[int] = timesteps * torch.ones(_A ,dtype=timesteps.dtype ,device=timesteps.device ) _lowerCAmelCase : Dict = self.time_proj(_A ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. _lowerCAmelCase : Any = timesteps_projected.to(dtype=self.dtype ) _lowerCAmelCase : Optional[Any] = self.time_embedding(_A ) if self.embedding_proj_norm is not None: _lowerCAmelCase : int = self.embedding_proj_norm(_A ) _lowerCAmelCase : str = self.embedding_proj(_A ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: _lowerCAmelCase : str = self.encoder_hidden_states_proj(_A ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError('`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set' ) _lowerCAmelCase : Any = self.proj_in(_A ) _lowerCAmelCase : Dict = self.positional_embedding.to(hidden_states.dtype ) _lowerCAmelCase : List[Any] = [] _lowerCAmelCase : Optional[Any] = 0 if encoder_hidden_states is not None: additional_embeds.append(_A ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: _lowerCAmelCase : int = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: _lowerCAmelCase : Any = hidden_states[:, None, :] _lowerCAmelCase : int = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: _lowerCAmelCase : List[str] = self.prd_embedding.to(hidden_states.dtype ).expand(_A ,-1 ,-1 ) additional_embeds.append(_A ) _lowerCAmelCase : List[str] = torch.cat( _A ,dim=1 ,) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens _lowerCAmelCase : Tuple = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: _lowerCAmelCase : Any = F.pad( _A ,( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) ,value=0.0 ,) _lowerCAmelCase : int = hidden_states + positional_embeddings if attention_mask is not None: _lowerCAmelCase : Optional[Any] = (1 - attention_mask.to(hidden_states.dtype )) * -1_0_0_0_0.0 _lowerCAmelCase : Union[str, Any] = F.pad(_A ,(0, self.additional_embeddings) ,value=0.0 ) _lowerCAmelCase : Tuple = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) _lowerCAmelCase : Optional[Any] = attention_mask.repeat_interleave(self.config.num_attention_heads ,dim=0 ) if self.norm_in is not None: _lowerCAmelCase : Any = self.norm_in(_A ) for block in self.transformer_blocks: _lowerCAmelCase : int = block(_A ,attention_mask=_A ) _lowerCAmelCase : Union[str, Any] = self.norm_out(_A ) if self.prd_embedding is not None: _lowerCAmelCase : Optional[int] = hidden_states[:, -1] else: _lowerCAmelCase : Any = hidden_states[:, additional_embeddings_len:] _lowerCAmelCase : Optional[int] = self.proj_to_clip_embeddings(_A ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=_A ) def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
16
1
"""simple docstring""" from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
65
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
1
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class lowercase__ ( __A , unittest.TestCase ): __UpperCamelCase = CpmAntTokenizer __UpperCamelCase = False def UpperCAmelCase__ ( self ): super().setUp() lowerCAmelCase_ : Optional[int] = [ """<d>""", """</d>""", """<s>""", """</s>""", """</_>""", """<unk>""", """<pad>""", """</n>""", """我""", """是""", """C""", """P""", """M""", """A""", """n""", """t""", ] lowerCAmelCase_ : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) @tooslow def UpperCAmelCase__ ( self ): lowerCAmelCase_ : List[Any] = CpmAntTokenizer.from_pretrained("""openbmb/cpm-ant-10b""" ) lowerCAmelCase_ : str = """今天天气真好!""" lowerCAmelCase_ : List[str] = ["""今天""", """天气""", """真""", """好""", """!"""] lowerCAmelCase_ : Dict = tokenizer.tokenize(_lowercase ) self.assertListEqual(_lowercase , _lowercase ) lowerCAmelCase_ : List[str] = """今天天气真好!""" lowerCAmelCase_ : Dict = [tokenizer.bos_token] + tokens lowerCAmelCase_ : List[str] = [6, 9_802, 14_962, 2_082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowercase ) , _lowercase ) lowerCAmelCase_ : str = tokenizer.decode(_lowercase ) self.assertEqual(_lowercase , _lowercase )
440
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : int = {"""configuration_plbart""": ["""PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PLBartConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = ["""PLBartTokenizer"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Union[str, Any] = [ """PLBART_PRETRAINED_MODEL_ARCHIVE_LIST""", """PLBartForCausalLM""", """PLBartForConditionalGeneration""", """PLBartForSequenceClassification""", """PLBartModel""", """PLBartPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys UpperCAmelCase_ : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
440
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor __magic_name__ = logging.get_logger(__name__) class _lowerCAmelCase ( lowerCamelCase ): def __init__( self , *a_ , **a_ ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , a_ , ) super().__init__(*a_ , **a_ )
657
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" return 10 - x * x def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" if equation(UpperCamelCase__ ) * equation(UpperCamelCase__ ) >= 0: raise ValueError("Wrong space!" ) _UpperCAmelCase = a while (b - a) >= 0.01: # Find middle point _UpperCAmelCase = (a + b) / 2 # Check if middle point is root if equation(UpperCamelCase__ ) == 0.0: break # Decide the side to repeat the steps if equation(UpperCamelCase__ ) * equation(UpperCamelCase__ ) < 0: _UpperCAmelCase = c else: _UpperCAmelCase = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
657
1
"""simple docstring""" from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) ): """simple docstring""" A_ : Union[str, Any] = tau * frequency / samplerate A_ : List[str] = sin(_UpperCAmelCase ) A_ : Union[str, Any] = cos(_UpperCAmelCase ) A_ : int = _sin / (2 * q_factor) A_ : Union[str, Any] = (1 - _cos) / 2 A_ : Tuple = 1 - _cos A_ : List[Any] = 1 + alpha A_ : List[Any] = -2 * _cos A_ : Union[str, Any] = 1 - alpha A_ : Optional[int] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) ): """simple docstring""" A_ : Dict = tau * frequency / samplerate A_ : Union[str, Any] = sin(_UpperCAmelCase ) A_ : int = cos(_UpperCAmelCase ) A_ : Union[str, Any] = _sin / (2 * q_factor) A_ : List[str] = (1 + _cos) / 2 A_ : int = -1 - _cos A_ : Union[str, Any] = 1 + alpha A_ : Tuple = -2 * _cos A_ : Any = 1 - alpha A_ : List[str] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) ): """simple docstring""" A_ : Tuple = tau * frequency / samplerate A_ : Optional[Any] = sin(_UpperCAmelCase ) A_ : str = cos(_UpperCAmelCase ) A_ : int = _sin / (2 * q_factor) A_ : Optional[int] = _sin / 2 A_ : Tuple = 0 A_ : Optional[Any] = -ba A_ : Dict = 1 + alpha A_ : List[Any] = -2 * _cos A_ : str = 1 - alpha A_ : Dict = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) ): """simple docstring""" A_ : str = tau * frequency / samplerate A_ : Dict = sin(_UpperCAmelCase ) A_ : Optional[Any] = cos(_UpperCAmelCase ) A_ : int = _sin / (2 * q_factor) A_ : Union[str, Any] = 1 - alpha A_ : List[str] = -2 * _cos A_ : Any = 1 + alpha A_ : Tuple = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) , ): """simple docstring""" A_ : str = tau * frequency / samplerate A_ : str = sin(_UpperCAmelCase ) A_ : List[Any] = cos(_UpperCAmelCase ) A_ : int = _sin / (2 * q_factor) A_ : Tuple = 10 ** (gain_db / 40) A_ : Any = 1 + alpha * big_a A_ : Union[str, Any] = -2 * _cos A_ : Tuple = 1 - alpha * big_a A_ : Optional[int] = 1 + alpha / big_a A_ : Any = -2 * _cos A_ : Optional[int] = 1 - alpha / big_a A_ : Dict = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) , ): """simple docstring""" A_ : List[str] = tau * frequency / samplerate A_ : Any = sin(_UpperCAmelCase ) A_ : Dict = cos(_UpperCAmelCase ) A_ : Dict = _sin / (2 * q_factor) A_ : Dict = 10 ** (gain_db / 40) A_ : Union[str, Any] = (big_a + 1) - (big_a - 1) * _cos A_ : Any = (big_a + 1) + (big_a - 1) * _cos A_ : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos A_ : List[Any] = (big_a - 1) + (big_a + 1) * _cos A_ : str = 2 * sqrt(_UpperCAmelCase ) * alpha A_ : Optional[int] = big_a * (pmc + aaa) A_ : Optional[int] = 2 * big_a * mpc A_ : Tuple = big_a * (pmc - aaa) A_ : Tuple = ppmc + aaa A_ : Optional[int] = -2 * pmpc A_ : Any = ppmc - aaa A_ : List[str] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def UpperCAmelCase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1 / sqrt(2 ) , ): """simple docstring""" A_ : str = tau * frequency / samplerate A_ : str = sin(_UpperCAmelCase ) A_ : Tuple = cos(_UpperCAmelCase ) A_ : Optional[Any] = _sin / (2 * q_factor) A_ : Tuple = 10 ** (gain_db / 40) A_ : Optional[int] = (big_a + 1) - (big_a - 1) * _cos A_ : Dict = (big_a + 1) + (big_a - 1) * _cos A_ : List[str] = (big_a - 1) - (big_a + 1) * _cos A_ : Union[str, Any] = (big_a - 1) + (big_a + 1) * _cos A_ : List[str] = 2 * sqrt(_UpperCAmelCase ) * alpha A_ : List[Any] = big_a * (ppmc + aaa) A_ : Union[str, Any] = -2 * big_a * pmpc A_ : List[Any] = big_a * (ppmc - aaa) A_ : Optional[Any] = pmc + aaa A_ : str = 2 * mpc A_ : Union[str, Any] = pmc - aaa A_ : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
713
"""simple docstring""" import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowerCamelCase_ : int = logging.get_logger(__name__) class _UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *snake_case_ , **snake_case_ ): """simple docstring""" warnings.warn( 'The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use LayoutLMv2ImageProcessor instead.' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
302
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """ut/deta""": """https://huggingface.co/ut/deta/resolve/main/config.json""", } class lowerCamelCase_ ( lowercase ): __lowercase : Optional[Any] = "deta" __lowercase : List[Any] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , lowerCamelCase_=None , lowerCamelCase_=9_00 , lowerCamelCase_=20_48 , lowerCamelCase_=6 , lowerCamelCase_=20_48 , lowerCamelCase_=8 , lowerCamelCase_=6 , lowerCamelCase_=10_24 , lowerCamelCase_=8 , lowerCamelCase_=0.0 , lowerCamelCase_=True , lowerCamelCase_="relu" , lowerCamelCase_=2_56 , lowerCamelCase_=0.1 , lowerCamelCase_=0.0 , lowerCamelCase_=0.0 , lowerCamelCase_=0.02 , lowerCamelCase_=1.0 , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_="sine" , lowerCamelCase_=5 , lowerCamelCase_=4 , lowerCamelCase_=4 , lowerCamelCase_=True , lowerCamelCase_=3_00 , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=1 , lowerCamelCase_=5 , lowerCamelCase_=2 , lowerCamelCase_=1 , lowerCamelCase_=1 , lowerCamelCase_=5 , lowerCamelCase_=2 , lowerCamelCase_=0.1 , lowerCamelCase_=0.25 , **lowerCamelCase_ , ) -> str: """simple docstring""" if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) _UpperCamelCase = CONFIG_MAPPING["resnet"](out_features=["stage2", "stage3", "stage4"] ) else: if isinstance(lowerCamelCase_ , lowerCamelCase_ ): _UpperCamelCase = backbone_config.pop("model_type" ) _UpperCamelCase = CONFIG_MAPPING[backbone_model_type] _UpperCamelCase = config_class.from_dict(lowerCamelCase_ ) _UpperCamelCase = backbone_config _UpperCamelCase = num_queries _UpperCamelCase = max_position_embeddings _UpperCamelCase = d_model _UpperCamelCase = encoder_ffn_dim _UpperCamelCase = encoder_layers _UpperCamelCase = encoder_attention_heads _UpperCamelCase = decoder_ffn_dim _UpperCamelCase = decoder_layers _UpperCamelCase = decoder_attention_heads _UpperCamelCase = dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = activation_function _UpperCamelCase = init_std _UpperCamelCase = init_xavier_std _UpperCamelCase = encoder_layerdrop _UpperCamelCase = auxiliary_loss _UpperCamelCase = position_embedding_type # deformable attributes _UpperCamelCase = num_feature_levels _UpperCamelCase = encoder_n_points _UpperCamelCase = decoder_n_points _UpperCamelCase = two_stage _UpperCamelCase = two_stage_num_proposals _UpperCamelCase = with_box_refine _UpperCamelCase = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("If two_stage is True, with_box_refine must be True." ) # Hungarian matcher _UpperCamelCase = class_cost _UpperCamelCase = bbox_cost _UpperCamelCase = giou_cost # Loss coefficients _UpperCamelCase = mask_loss_coefficient _UpperCamelCase = dice_loss_coefficient _UpperCamelCase = bbox_loss_coefficient _UpperCamelCase = giou_loss_coefficient _UpperCamelCase = eos_coefficient _UpperCamelCase = focal_alpha super().__init__(is_encoder_decoder=lowerCamelCase_ , **lowerCamelCase_ ) @property def lowercase ( self ) -> int: """simple docstring""" return self.encoder_attention_heads @property def lowercase ( self ) -> int: """simple docstring""" return self.d_model def lowercase ( self ) -> Optional[int]: """simple docstring""" _UpperCamelCase = copy.deepcopy(self.__dict__ ) _UpperCamelCase = self.backbone_config.to_dict() _UpperCamelCase = self.__class__.model_type return output
147
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class lowerCamelCase_ : def __init__( self , lowerCamelCase_ , lowerCamelCase_=13 , lowerCamelCase_=7 , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=99 , lowerCamelCase_=32 , lowerCamelCase_=2 , lowerCamelCase_=4 , lowerCamelCase_=37 , lowerCamelCase_="gelu" , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=5_12 , lowerCamelCase_=16 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=3 , lowerCamelCase_=4 , lowerCamelCase_=None , ) -> Optional[int]: """simple docstring""" _UpperCamelCase = parent _UpperCamelCase = 13 _UpperCamelCase = 7 _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = 99 _UpperCamelCase = 3_84 _UpperCamelCase = 2 _UpperCamelCase = 4 _UpperCamelCase = 37 _UpperCamelCase = "gelu" _UpperCamelCase = 0.1 _UpperCamelCase = 0.1 _UpperCamelCase = 5_12 _UpperCamelCase = 16 _UpperCamelCase = 2 _UpperCamelCase = 0.02 _UpperCamelCase = 3 _UpperCamelCase = 4 _UpperCamelCase = 1_28 _UpperCamelCase = 2 _UpperCamelCase = 9 _UpperCamelCase = 1 _UpperCamelCase = None def lowercase ( self ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _UpperCamelCase = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=lowerCamelCase_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[Any]: """simple docstring""" _UpperCamelCase = TFConvBertModel(config=lowerCamelCase_ ) _UpperCamelCase = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _UpperCamelCase = [input_ids, input_mask] _UpperCamelCase = model(lowerCamelCase_ ) _UpperCamelCase = model(lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = TFConvBertForMaskedLM(config=lowerCamelCase_ ) _UpperCamelCase = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _UpperCamelCase = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: """simple docstring""" _UpperCamelCase = self.num_labels _UpperCamelCase = TFConvBertForSequenceClassification(config=lowerCamelCase_ ) _UpperCamelCase = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _UpperCamelCase = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: """simple docstring""" _UpperCamelCase = self.num_choices _UpperCamelCase = TFConvBertForMultipleChoice(config=lowerCamelCase_ ) _UpperCamelCase = tf.tile(tf.expand_dims(lowerCamelCase_ , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase = tf.tile(tf.expand_dims(lowerCamelCase_ , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase = tf.tile(tf.expand_dims(lowerCamelCase_ , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _UpperCamelCase = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[Any]: """simple docstring""" _UpperCamelCase = self.num_labels _UpperCamelCase = TFConvBertForTokenClassification(config=lowerCamelCase_ ) _UpperCamelCase = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _UpperCamelCase = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: """simple docstring""" _UpperCamelCase = TFConvBertForQuestionAnswering(config=lowerCamelCase_ ) _UpperCamelCase = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _UpperCamelCase = model(lowerCamelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase ( self ) -> str: """simple docstring""" _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class lowerCamelCase_ ( lowercase , lowercase , unittest.TestCase ): __lowercase : int = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : int = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Any = False __lowercase : List[str] = False __lowercase : List[Any] = False def lowercase ( self ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = TFConvBertModelTester(self ) _UpperCamelCase = ConfigTester(self , config_class=lowerCamelCase_ , hidden_size=37 ) def lowercase ( self ) -> int: """simple docstring""" self.config_tester.run_common_tests() def lowercase ( self ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def lowercase ( self ) -> Any: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase_ ) def lowercase ( self ) -> int: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCamelCase_ ) def lowercase ( self ) -> Dict: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase_ ) def lowercase ( self ) -> Tuple: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase_ ) def lowercase ( self ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase_ ) @slow def lowercase ( self ) -> Dict: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = True _UpperCamelCase = True if hasattr(lowerCamelCase_ , "use_cache" ): _UpperCamelCase = True _UpperCamelCase = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _UpperCamelCase = getattr(self.model_tester , "key_length" , lowerCamelCase_ ) for model_class in self.all_model_classes: _UpperCamelCase = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) _UpperCamelCase = model_class(lowerCamelCase_ ) _UpperCamelCase = len(model(lowerCamelCase_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase_ , saved_model=lowerCamelCase_ ) _UpperCamelCase = os.path.join(lowerCamelCase_ , "saved_model" , "1" ) _UpperCamelCase = tf.keras.models.load_model(lowerCamelCase_ ) _UpperCamelCase = model(lowerCamelCase_ ) if self.is_encoder_decoder: _UpperCamelCase = outputs["encoder_hidden_states"] _UpperCamelCase = outputs["encoder_attentions"] else: _UpperCamelCase = outputs["hidden_states"] _UpperCamelCase = outputs["attentions"] self.assertEqual(len(lowerCamelCase_ ) , lowerCamelCase_ ) _UpperCamelCase = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(lowerCamelCase_ ) , lowerCamelCase_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(lowerCamelCase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def lowercase ( self ) -> List[Any]: """simple docstring""" _UpperCamelCase = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(lowerCamelCase_ ) def lowercase ( self ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = True _UpperCamelCase = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) _UpperCamelCase = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _UpperCamelCase = getattr(self.model_tester , "key_length" , lowerCamelCase_ ) _UpperCamelCase = getattr(self.model_tester , "key_length" , lowerCamelCase_ ) def check_decoder_attentions_output(lowerCamelCase_ ): _UpperCamelCase = len(lowerCamelCase_ ) self.assertEqual(out_len % 2 , 0 ) _UpperCamelCase = outputs.decoder_attentions self.assertEqual(len(lowerCamelCase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(lowerCamelCase_ ): _UpperCamelCase = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(lowerCamelCase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: _UpperCamelCase = True _UpperCamelCase = False _UpperCamelCase = model_class(lowerCamelCase_ ) _UpperCamelCase = model(self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) _UpperCamelCase = len(lowerCamelCase_ ) self.assertEqual(config.output_hidden_states , lowerCamelCase_ ) check_encoder_attentions_output(lowerCamelCase_ ) if self.is_encoder_decoder: _UpperCamelCase = model_class(lowerCamelCase_ ) _UpperCamelCase = model(self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) self.assertEqual(config.output_hidden_states , lowerCamelCase_ ) check_decoder_attentions_output(lowerCamelCase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _UpperCamelCase = True _UpperCamelCase = model_class(lowerCamelCase_ ) _UpperCamelCase = model(self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) self.assertEqual(config.output_hidden_states , lowerCamelCase_ ) check_encoder_attentions_output(lowerCamelCase_ ) # Check attention is always last and order is fine _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = model_class(lowerCamelCase_ ) _UpperCamelCase = model(self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowerCamelCase_ ) ) self.assertEqual(model.config.output_hidden_states , lowerCamelCase_ ) check_encoder_attentions_output(lowerCamelCase_ ) @require_tf class lowerCamelCase_ ( unittest.TestCase ): @slow def lowercase ( self ) -> int: """simple docstring""" _UpperCamelCase = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) _UpperCamelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) _UpperCamelCase = model(lowerCamelCase_ )[0] _UpperCamelCase = [1, 6, 7_68] self.assertEqual(output.shape , lowerCamelCase_ ) _UpperCamelCase = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowerCamelCase_ , atol=1E-4 )
147
1
from __future__ import annotations def lowercase__ ( __A: Union[str, Any] ): '''simple docstring''' for i in range(1 ,len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 ,len(__SCREAMING_SNAKE_CASE ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 ,len(__SCREAMING_SNAKE_CASE ) ): for j in range(1 ,len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] ,matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
704
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __lowerCamelCase : int = {'''configuration_swin''': ['''SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwinConfig''', '''SwinOnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''SWIN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwinForImageClassification''', '''SwinForMaskedImageModeling''', '''SwinModel''', '''SwinPreTrainedModel''', '''SwinBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[int] = [ '''TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFSwinForImageClassification''', '''TFSwinForMaskedImageModeling''', '''TFSwinModel''', '''TFSwinPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys __lowerCamelCase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
501
0
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( """The `inpainting.py` script is outdated. Please use directly `from diffusers import""" """ StableDiffusionInpaintPipeline` instead.""" )
175
def _lowercase ( __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : List[Any] , __UpperCamelCase : List[Any] ): if height >= 1: move_tower(height - 1 , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) move_disk(__UpperCamelCase , __UpperCamelCase ) move_tower(height - 1 , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def _lowercase ( __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int] ): print("""moving disk from""" , __UpperCamelCase , """to""" , __UpperCamelCase ) def _lowercase ( ): snake_case__ = int(input("""Height of hanoi: """ ).strip() ) move_tower(__UpperCamelCase , """A""" , """B""" , """C""" ) if __name__ == "__main__": main()
214
0
from collections.abc import Generator from math import sin def UpperCAmelCase_ ( _A ): '''simple docstring''' if len(_A ) != 32: raise ValueError('''Input must be of length 32''' ) SCREAMING_SNAKE_CASE__ = B'''''' for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCAmelCase_ ( _A ): '''simple docstring''' if i < 0: raise ValueError('''Input must be non-negative''' ) SCREAMING_SNAKE_CASE__ = format(_A , '''08x''' )[-8:] SCREAMING_SNAKE_CASE__ = B'''''' for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode('''utf-8''' ) return little_endian_hex def UpperCAmelCase_ ( _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = B'''''' for char in message: bit_string += format(_A , '''08b''' ).encode('''utf-8''' ) SCREAMING_SNAKE_CASE__ = format(len(_A ) , '''064b''' ).encode('''utf-8''' ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(_A ) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def UpperCAmelCase_ ( _A ): '''simple docstring''' if len(_A ) % 5_12 != 0: raise ValueError('''Input must have length that\'s a multiple of 512''' ) for pos in range(0 , len(_A ) , 5_12 ): SCREAMING_SNAKE_CASE__ = bit_string[pos : pos + 5_12] SCREAMING_SNAKE_CASE__ = [] for i in range(0 , 5_12 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def UpperCAmelCase_ ( _A ): '''simple docstring''' if i < 0: raise ValueError('''Input must be non-negative''' ) SCREAMING_SNAKE_CASE__ = format(_A , '''032b''' ) SCREAMING_SNAKE_CASE__ = '''''' for c in i_str: new_str += "1" if c == "0" else "0" return int(_A , 2 ) def UpperCAmelCase_ ( _A , _A ): '''simple docstring''' return (a + b) % 2**32 def UpperCAmelCase_ ( _A , _A ): '''simple docstring''' if i < 0: raise ValueError('''Input must be non-negative''' ) if shift < 0: raise ValueError('''Shift must be non-negative''' ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCAmelCase_ ( _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = preprocess(_A ) SCREAMING_SNAKE_CASE__ = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states SCREAMING_SNAKE_CASE__ = 0x6_7_4_5_2_3_0_1 SCREAMING_SNAKE_CASE__ = 0xe_f_c_d_a_b_8_9 SCREAMING_SNAKE_CASE__ = 0x9_8_b_a_d_c_f_e SCREAMING_SNAKE_CASE__ = 0x1_0_3_2_5_4_7_6 SCREAMING_SNAKE_CASE__ = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(_A ): SCREAMING_SNAKE_CASE__ = aa SCREAMING_SNAKE_CASE__ = ba SCREAMING_SNAKE_CASE__ = ca SCREAMING_SNAKE_CASE__ = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f SCREAMING_SNAKE_CASE__ = d ^ (b & (c ^ d)) SCREAMING_SNAKE_CASE__ = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f SCREAMING_SNAKE_CASE__ = c ^ (d & (b ^ c)) SCREAMING_SNAKE_CASE__ = (5 * i + 1) % 16 elif i <= 47: SCREAMING_SNAKE_CASE__ = b ^ c ^ d SCREAMING_SNAKE_CASE__ = (3 * i + 5) % 16 else: SCREAMING_SNAKE_CASE__ = c ^ (b | not_aa(_A )) SCREAMING_SNAKE_CASE__ = (7 * i) % 16 SCREAMING_SNAKE_CASE__ = (f + a + added_consts[i] + block_words[g]) % 2**32 SCREAMING_SNAKE_CASE__ = d SCREAMING_SNAKE_CASE__ = c SCREAMING_SNAKE_CASE__ = b SCREAMING_SNAKE_CASE__ = sum_aa(_A , left_rotate_aa(_A , shift_amounts[i] ) ) # Add hashed chunk to running total SCREAMING_SNAKE_CASE__ = sum_aa(_A , _A ) SCREAMING_SNAKE_CASE__ = sum_aa(_A , _A ) SCREAMING_SNAKE_CASE__ = sum_aa(_A , _A ) SCREAMING_SNAKE_CASE__ = sum_aa(_A , _A ) SCREAMING_SNAKE_CASE__ = reformat_hex(_A ) + reformat_hex(_A ) + reformat_hex(_A ) + reformat_hex(_A ) return digest if __name__ == "__main__": import doctest doctest.testmod()
472
from dataclasses import dataclass, field from typing import Optional @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default="codeparrot/codeparrot" , metadata={"help": "Model name or path of model to be trained."} ) a = field( default="./" , metadata={"help": "Save dir where model repo is cloned and models updates are saved to."} ) a = field( default="codeparrot/codeparrot-clean-train" , metadata={"help": "Name or path of training dataset."} ) a = field( default="codeparrot/codeparrot-clean-valid" , metadata={"help": "Name or path of validation dataset."} ) a = field(default=2 , metadata={"help": "Batch size for training."} ) a = field(default=2 , metadata={"help": "Batch size for evaluation."} ) a = field(default=0.1 , metadata={"help": "Value of weight decay."} ) a = field( default=1_00_00 , metadata={"help": "Size of buffer used to shuffle streaming dataset."} ) a = field(default=2e-4 , metadata={"help": "Learning rate fo training."} ) a = field(default="cosine" , metadata={"help": "Learning rate."} ) a = field( default=7_50 , metadata={"help": "Number of warmup steps in the learning rate schedule."} ) a = field( default=16 , metadata={"help": "Number of gradient accumulation steps."} ) a = field( default=A__ , metadata={"help": "Use gradient checkpointing to reduce memory footprint."} ) a = field(default=5_00_00 , metadata={"help": "Maximum number of training steps."} ) a = field( default=-1 , metadata={"help": "Maximum number of evaluation steps. If -1 the full dataset is evaluated."} ) a = field(default=10_24 , metadata={"help": "Sequence lengths used for training."} ) a = field(default=1 , metadata={"help": "Training seed."} ) a = field( default=10_24 , metadata={"help": "Interval to save checkpoints. Measured as number of forward passes not training steps."} , ) a = field( default=A__ , metadata={"help": "States path if the training should continue from a checkpoint folder."} ) a = field(default=A__ , metadata={"help": "If True the data is pretokenized."} ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default="codeparrot/codeparrot" , metadata={"help": "Model name or path of model to be evaluated."} ) a = field( default="codeparrot/codeparrot-clean-valid" , metadata={"help": "Name or path of validation dataset."} ) a = field(default=2 , metadata={"help": "Batch size used for evaluation."} ) a = field( default=-1 , metadata={"help": "Maximum number of evaluation steps. If -1 the full dataset is evaluated."} ) a = field(default=10_24 , metadata={"help": "Length of sequences to be evaluated."} ) a = field(default=1 , metadata={"help": "Random seed used for evaluation."} ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default="codeparrot/codeparrot" , metadata={"help": "Model name or path of model to be evaluated."} ) a = field(default=A__ , metadata={"help": "Number of workers used for code evaluation."} ) a = field( default=A__ , metadata={"help": "The number of human-eval tasks to run. If not included all tasks are evaluated."} , ) a = field( default=A__ , metadata={"help": "Sample from the language model's output distribution."} ) a = field(default=0.2 , metadata={"help": "Sampling temperature used for generation."} ) a = field(default=2_56 , metadata={"help": "Maximum number of newly generated tokens."} ) a = field(default=0 , metadata={"help": "Top-k parameter used for generation."} ) a = field(default=0.9_5 , metadata={"help": "Top-p parameter used for nucleus sampling."} ) a = field(default=10 , metadata={"help": "Number of generations to run in parallel."} ) a = field( default=2_00 , metadata={"help": "Number of completions to generate for each sample."} ) a = field(default=1 , metadata={"help": "Random seed used for evaluation."} ) a = field( default="eval_results.json" , metadata={"help": "Random seed used for evaluation."} ) a = field( default="0" , metadata={"help": "Allow `code_eval` to execute Python code on machine"} ) a = field( default=-1 , metadata={ "help": ( "Determine which device to run the `text-generation` Pipeline on. -1 is CPU and any zero or positive" " number corresponds to which GPU device id to run on." ) } , ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default=A__ , metadata={ "help": "The number of CPU cores to use for parallel preprocessing. Default uses the maximum available." } , ) a = field( default="transformersbook/codeparrot" , metadata={"help": "Folder or name of dataset to process."} ) a = field( default="codeparrot-clean" , metadata={"help": "Folder to save processed processed dataset."} ) a = field( default=10_00_00 , metadata={"help": "Number of files to save per JSON output file."} ) a = field(default="content" , metadata={"help": "Column containing text data to process."} ) a = field( default=10_00 , metadata={"help": "Maximum line length in file, otherwise file is filtered."} ) a = field( default=1_00 , metadata={"help": "Maximum mean line length in file, otherwise file is filtered."} ) a = field( default=0.2_5 , metadata={"help": "Maximum fraction of non-alphanumeric characters, otherwise file is filtered."} ) a = field( default=1.5 , metadata={"help": "Minimum character token ratio for the file, otherwise file is filtered."} ) a = field( default=0.7 , metadata={"help": "Probability for filtering config, test and uncommon files."} ) a = field( default="codeparrot/codeparrot" , metadata={"help": "Name or path to the tokenizer."} , ) a = field( default=A__ , metadata={"help": "If True, near-duplicate samples are removed."} ) a = field( default=0.8_5 , metadata={"help": "Jaccard threshold for near-duplicate samples."} ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default="gpt2" , metadata={"help": "Base tokenizer to build new tokenizer from."} ) a = field( default="transformersbook/codeparrot-train" , metadata={"help": "Dataset to train tokenizer on."} ) a = field(default="content" , metadata={"help": "Column containing text data to process."} ) a = field(default=20_00_00 , metadata={"help": "Number of examples to train tokenizer on."} ) a = field( default=3_27_68 , metadata={"help": "Number of examples to train the tokenizer on."} ) a = field(default="codeparrot" , metadata={"help": "Name of new tokenizer."} ) a = field(default=A__ , metadata={"help": "Push saved tokenizer to the hub."} ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default="codeparrot/codeparrot" , metadata={"help": "Name or path to the tokenizer."} ) a = field( default="codeparrot/codeparrot-clean-train" , metadata={"help": "Name or path to the dataset to pretokenize."} ) a = field( default="tokenized-codeparrot-train" , metadata={"help": "Repo name of the pretokenized data."} ) a = field(default=A__ , metadata={"help": "Number of workers used for code evaluation."} ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( default="gpt2-large" , metadata={"help": "Configuration to use for model initialization."} ) a = field( default="codeparrot/codeparrot" , metadata={"help": "Tokenizer attached to model."} ) a = field(default="codeparrot" , metadata={"help": "Name of the created model."} ) a = field(default=A__ , metadata={"help": "Push saved tokenizer to the hub."} )
472
1
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": UpperCAmelCase : int = pd.read_csv("sample_data.csv", header=None) UpperCAmelCase : Optional[Any] = df.shape[:1][0] # If you're using some other dataset input the target column UpperCAmelCase : List[Any] = df.iloc[:, 1:2] UpperCAmelCase : Optional[Any] = actual_data.values.reshape(len_data, 1) UpperCAmelCase : Dict = MinMaxScaler().fit_transform(actual_data) UpperCAmelCase : List[str] = 10 UpperCAmelCase : Optional[int] = 5 UpperCAmelCase : Optional[Any] = 20 UpperCAmelCase : Union[str, Any] = len_data - periods * look_back UpperCAmelCase : int = actual_data[:division] UpperCAmelCase : str = actual_data[division - look_back :] UpperCAmelCase, UpperCAmelCase : int = [], [] UpperCAmelCase, UpperCAmelCase : List[Any] = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) UpperCAmelCase : Optional[int] = np.array(train_x) UpperCAmelCase : Union[str, Any] = np.array(test_x) UpperCAmelCase : int = np.array([list(i.ravel()) for i in train_y]) UpperCAmelCase : Any = np.array([list(i.ravel()) for i in test_y]) UpperCAmelCase : List[Any] = Sequential() model.add(LSTM(1_28, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(1_28, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") UpperCAmelCase : List[str] = model.fit( x_train, y_train, epochs=1_50, verbose=1, shuffle=True, batch_size=4 ) UpperCAmelCase : List[Any] = model.predict(x_test)
457
import os from datetime import datetime as dt from github import Github UpperCAmelCase : str = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def __lowerCamelCase ( ): '''simple docstring''' lowerCamelCase = Github(os.environ["""GITHUB_TOKEN"""] ) lowerCamelCase = g.get_repo("""huggingface/diffusers""" ) lowerCamelCase = repo.get_issues(state="""open""" ) for issue in open_issues: lowerCamelCase = sorted(issue.get_comments() , key=lambda lowerCamelCase__ : i.created_at , reverse=lowerCamelCase__ ) lowerCamelCase = comments[0] if len(lowerCamelCase__ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state="""closed""" ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state="""open""" ) issue.remove_from_labels("""stale""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) issue.add_to_labels("""stale""" ) if __name__ == "__main__": main()
457
1
def UpperCamelCase_( _A :str )-> int: UpperCamelCase__ = hex_num.strip() if not hex_num: raise ValueError("No value was passed to the function" ) UpperCamelCase__ = hex_num[0] == "-" if is_negative: UpperCamelCase__ = hex_num[1:] try: UpperCamelCase__ = int(_A , 16 ) except ValueError: raise ValueError("Invalid value was passed to the function" ) UpperCamelCase__ = "" while int_num > 0: UpperCamelCase__ = str(int_num % 2 ) + bin_str int_num >>= 1 return int(("-" + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
185
from __future__ import annotations def UpperCamelCase_( _A :list[int] , _A :int )-> list[int]: UpperCamelCase__ = 0 UpperCamelCase__ = len(_A ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: UpperCamelCase__ = i + 1 else: UpperCamelCase__ = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f'''{two_pointer([2, 7, 1_1, 1_5], 9) = }''')
185
1
"""simple docstring""" import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" def __init__( self :Tuple , __lowercase :Union[str, Any] , __lowercase :Any=13 , __lowercase :Any=30 , __lowercase :Tuple=2 , __lowercase :List[Any]=3 , __lowercase :Any=True , __lowercase :List[Any]=True , __lowercase :List[Any]=32 , __lowercase :Optional[int]=5 , __lowercase :Tuple=4 , __lowercase :Union[str, Any]=37 , __lowercase :List[Any]="gelu" , __lowercase :Optional[int]=0.1 , __lowercase :Optional[Any]=0.1 , __lowercase :Any=10 , __lowercase :Union[str, Any]=0.02 , ): __lowerCamelCase : Dict =parent __lowerCamelCase : str =batch_size __lowerCamelCase : Dict =image_size __lowerCamelCase : List[str] =patch_size __lowerCamelCase : List[Any] =num_channels __lowerCamelCase : Union[str, Any] =is_training __lowerCamelCase : Tuple =use_labels __lowerCamelCase : int =hidden_size __lowerCamelCase : Union[str, Any] =num_hidden_layers __lowerCamelCase : Any =num_attention_heads __lowerCamelCase : Tuple =intermediate_size __lowerCamelCase : str =hidden_act __lowerCamelCase : List[Any] =hidden_dropout_prob __lowerCamelCase : Optional[int] =attention_probs_dropout_prob __lowerCamelCase : str =type_sequence_label_size __lowerCamelCase : int =initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCamelCase : List[Any] =(image_size // patch_size) ** 2 __lowerCamelCase : Optional[Any] =num_patches + 1 def __lowercase ( self :str ): __lowerCamelCase : str =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase : Optional[int] =ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowercase , initializer_range=self.initializer_range , ) return config, pixel_values def __lowercase ( self :List[str] , __lowercase :Tuple , __lowercase :Tuple ): __lowerCamelCase : List[Any] =FlaxViTModel(config=__lowercase ) __lowerCamelCase : Optional[Any] =model(__lowercase ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) __lowerCamelCase : Dict =(self.image_size, self.image_size) __lowerCamelCase : Dict =(self.patch_size, self.patch_size) __lowerCamelCase : List[str] =(image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, num_patches + 1, self.hidden_size) ) def __lowercase ( self :str , __lowercase :List[str] , __lowercase :Optional[int] ): __lowerCamelCase : Union[str, Any] =self.type_sequence_label_size __lowerCamelCase : Tuple =FlaxViTForImageClassification(config=__lowercase ) __lowerCamelCase : int =model(__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCamelCase : int =1 __lowerCamelCase : Optional[Any] =FlaxViTForImageClassification(__lowercase ) __lowerCamelCase : List[str] =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCamelCase : Dict =model(__lowercase ) def __lowercase ( self :Union[str, Any] ): __lowerCamelCase : str =self.prepare_config_and_inputs() ( ( __lowerCamelCase ) , ( __lowerCamelCase ) , ) : Dict =config_and_inputs __lowerCamelCase : str ={'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class SCREAMING_SNAKE_CASE_ ( snake_case__ , unittest.TestCase ): """simple docstring""" __snake_case : Any = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def __lowercase ( self :Optional[Any] ): __lowerCamelCase : Tuple =FlaxViTModelTester(self ) __lowerCamelCase : Tuple =ConfigTester(self , config_class=__lowercase , has_text_modality=__lowercase , hidden_size=37 ) def __lowercase ( self :Tuple ): self.config_tester.run_common_tests() def __lowercase ( self :Tuple ): __lowerCamelCase : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def __lowercase ( self :int ): __lowerCamelCase : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) def __lowercase ( self :str ): __lowerCamelCase , __lowerCamelCase : int =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase : Optional[int] =model_class(__lowercase ) __lowerCamelCase : str =inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase : int =[*signature.parameters.keys()] __lowerCamelCase : List[str] =['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowercase ) def __lowercase ( self :Optional[Any] ): __lowerCamelCase , __lowerCamelCase : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __lowerCamelCase : Any =self._prepare_for_class(__lowercase , __lowercase ) __lowerCamelCase : str =model_class(__lowercase ) @jax.jit def model_jitted(__lowercase :int , **__lowercase :Optional[int] ): return model(pixel_values=__lowercase , **__lowercase ) with self.subTest('''JIT Enabled''' ): __lowerCamelCase : Dict =model_jitted(**__lowercase ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): __lowerCamelCase : List[str] =model_jitted(**__lowercase ).to_tuple() self.assertEqual(len(__lowercase ) , len(__lowercase ) ) for jitted_output, output in zip(__lowercase , __lowercase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def __lowercase ( self :str ): for model_class_name in self.all_model_classes: __lowerCamelCase : Any =model_class_name.from_pretrained('''google/vit-base-patch16-224''' ) __lowerCamelCase : Optional[Any] =model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(__lowercase )
179
"""simple docstring""" from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE : str ): '''simple docstring''' if not is_accelerate_available(): return method __lowerCamelCase : Dict =version.parse(accelerate.__version__ ).base_version if version.parse(SCREAMING_SNAKE_CASE ) < version.parse('''0.17.0''' ): return method def wrapper(self : List[Any] , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : List[str] ): if hasattr(self , '''_hf_hook''' ) and hasattr(self._hf_hook , '''pre_forward''' ): self._hf_hook.pre_forward(self ) return method(self , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) return wrapper
179
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class __magic_name__ ( unittest.TestCase): @slow def _UpperCAmelCase ( self : Tuple ): UpperCAmelCase = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) UpperCAmelCase = torch.tensor([[0, 5_8_1, 1_0_2_6_9, 8_3, 9_9_9_4_2, 1_3_6, 6_0_7_4_2, 2_3, 7_0, 8_0_5_8_3, 1_8_2_7_6, 2]] ) # The dog is cute and lives in the garden house UpperCAmelCase = torch.Size((1, 1_2, 7_6_8) ) # batch_size, sequence_length, embedding_vector_dim UpperCAmelCase = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): UpperCAmelCase = model(__SCREAMING_SNAKE_CASE )["last_hidden_state"].detach() self.assertEqual(output.shape ,__SCREAMING_SNAKE_CASE ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] ,__SCREAMING_SNAKE_CASE ,atol=1e-3 ) ) @slow def _UpperCAmelCase ( self : int ): UpperCAmelCase = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) UpperCAmelCase = torch.tensor([[0, 5_8_1, 1_0_2_6_9, 8_3, 9_9_9_4_2, 1_3_6, 6_0_7_4_2, 2_3, 7_0, 8_0_5_8_3, 1_8_2_7_6, 2]] ) # The dog is cute and lives in the garden house UpperCAmelCase = torch.Size((1, 1_2, 1_0_2_4) ) # batch_size, sequence_length, embedding_vector_dim UpperCAmelCase = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): UpperCAmelCase = model(__SCREAMING_SNAKE_CASE )["last_hidden_state"].detach() self.assertEqual(output.shape ,__SCREAMING_SNAKE_CASE ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] ,__SCREAMING_SNAKE_CASE ,atol=1e-3 ) )
405
def __UpperCamelCase ( _lowerCAmelCase = 10 ): """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or n < 0: raise ValueError("Invalid input" ) UpperCAmelCase = 10**n UpperCAmelCase = 2_84_33 * (pow(2 , 7_83_04_57 , _lowerCAmelCase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f"{solution(10) = }")
405
1
'''simple docstring''' from __future__ import annotations def _lowerCAmelCase ( lowerCamelCase_ : int ): __lowercase = str(lowerCamelCase_ ) return len(lowerCamelCase_ ) == 9 and set(lowerCamelCase_ ) == set('''123456789''' ) def _lowerCAmelCase ( ): for base_num in range(9_9_9_9 , 4_9_9_9 , -1 ): __lowercase = 1_0_0_0_0_2 * base_num if is_9_pandigital(lowerCamelCase_ ): return candidate for base_num in range(3_3_3 , 9_9 , -1 ): __lowercase = 1_0_0_2_0_0_3 * base_num if is_9_pandigital(lowerCamelCase_ ): return candidate return None if __name__ == "__main__": print(f'''{solution() = }''')
502
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' a : int = "distilbert" a : Union[str, Any] = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__(self ,_lowerCamelCase=30522 ,_lowerCamelCase=512 ,_lowerCamelCase=False ,_lowerCamelCase=6 ,_lowerCamelCase=12 ,_lowerCamelCase=768 ,_lowerCamelCase=4 * 768 ,_lowerCamelCase=0.1 ,_lowerCamelCase=0.1 ,_lowerCamelCase="gelu" ,_lowerCamelCase=0.0_2 ,_lowerCamelCase=0.1 ,_lowerCamelCase=0.2 ,_lowerCamelCase=0 ,**_lowerCamelCase ,) -> Tuple: '''simple docstring''' __lowercase = vocab_size __lowercase = max_position_embeddings __lowercase = sinusoidal_pos_embds __lowercase = n_layers __lowercase = n_heads __lowercase = dim __lowercase = hidden_dim __lowercase = dropout __lowercase = attention_dropout __lowercase = activation __lowercase = initializer_range __lowercase = qa_dropout __lowercase = seq_classif_dropout super().__init__(**_lowerCamelCase ,pad_token_id=_lowerCamelCase ) class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' @property def _UpperCAmelCase (self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": __lowercase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __lowercase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
502
1
"""simple docstring""" import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model _snake_case = '''0.12''' # assumed parallelism: 8 if is_torch_available(): import torch def __snake_case ( SCREAMING_SNAKE_CASE: Tuple , SCREAMING_SNAKE_CASE: Union[str, Any] , SCREAMING_SNAKE_CASE: List[Any]=None ): """simple docstring""" if rng is None: _lowerCAmelCase = random.Random() _lowerCAmelCase = 1 for dim in shape: total_dims *= dim _lowerCAmelCase = [] for _ in range(SCREAMING_SNAKE_CASE ): values.append(rng.randint(0 , vocab_size - 1 ) ) _lowerCAmelCase = np.array(SCREAMING_SNAKE_CASE , dtype=jnp.intaa ).reshape(SCREAMING_SNAKE_CASE ) return output def __snake_case ( SCREAMING_SNAKE_CASE: List[str] , SCREAMING_SNAKE_CASE: Optional[int]=None ): """simple docstring""" _lowerCAmelCase = ids_tensor(SCREAMING_SNAKE_CASE , vocab_size=2 , rng=SCREAMING_SNAKE_CASE ) # make sure that at least one token is attended to for each batch _lowerCAmelCase = 1 return attn_mask @require_flax class _SCREAMING_SNAKE_CASE : '''simple docstring''' SCREAMING_SNAKE_CASE_: Tuple = None SCREAMING_SNAKE_CASE_: List[Any] = () def __lowerCamelCase ( self : str ) -> Optional[Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 _lowerCAmelCase = 2 _lowerCAmelCase = inputs['input_ids'].shape[-1] // 2 _lowerCAmelCase = inputs['input_ids'][:max_batch_size, :sequence_length] _lowerCAmelCase = jnp.ones_like(UpperCAmelCase_ ) _lowerCAmelCase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens _lowerCAmelCase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` _lowerCAmelCase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def __lowerCamelCase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = False _lowerCAmelCase = max_length _lowerCAmelCase = 0 for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning _lowerCAmelCase = getattr(UpperCAmelCase_ , UpperCAmelCase_ ) _lowerCAmelCase = pt_model_class(UpperCAmelCase_ ).eval() _lowerCAmelCase = load_flax_weights_in_pytorch_model(UpperCAmelCase_ , flax_model.params ) _lowerCAmelCase = flax_model.generate(UpperCAmelCase_ ).sequences _lowerCAmelCase = pt_model.generate(torch.tensor(UpperCAmelCase_ , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: _lowerCAmelCase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def __lowerCamelCase ( self : List[str] ) -> List[str]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = False _lowerCAmelCase = max_length for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : Any ) -> Optional[Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = True _lowerCAmelCase = max_length for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = False _lowerCAmelCase = max_length _lowerCAmelCase = 2 for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : str ) -> List[str]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = False _lowerCAmelCase = max_length _lowerCAmelCase = 2 _lowerCAmelCase = 2 for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def __lowerCamelCase ( self : int ) -> Any: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = True _lowerCAmelCase = max_length _lowerCAmelCase = 0.8 _lowerCAmelCase = 10 _lowerCAmelCase = 0.3 _lowerCAmelCase = 1 _lowerCAmelCase = 8 _lowerCAmelCase = 9 for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = max_length _lowerCAmelCase = 1 _lowerCAmelCase = 8 _lowerCAmelCase = 9 for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : List[Any] ) -> List[str]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() _lowerCAmelCase = max_length _lowerCAmelCase = 2 _lowerCAmelCase = 1 _lowerCAmelCase = 8 _lowerCAmelCase = 9 for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() # pad attention mask on the left _lowerCAmelCase = attention_mask.at[(0, 0)].set(0 ) _lowerCAmelCase = False _lowerCAmelCase = max_length for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() # pad attention mask on the left _lowerCAmelCase = attention_mask.at[(0, 0)].set(0 ) _lowerCAmelCase = True _lowerCAmelCase = max_length for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def __lowerCamelCase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = self._get_input_ids_and_config() # pad attention mask on the left _lowerCAmelCase = attention_mask.at[(0, 0)].set(0 ) _lowerCAmelCase = 2 _lowerCAmelCase = max_length for model_class in self.all_generative_model_classes: _lowerCAmelCase = model_class(UpperCAmelCase_ ) _lowerCAmelCase = model.generate(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase_ ) _lowerCAmelCase = jit(model.generate ) _lowerCAmelCase = jit_generate(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __lowerCamelCase ( self : Tuple ) -> str: """simple docstring""" _lowerCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-bert' ) _lowerCAmelCase = FlaxAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-bert-flax-only' ) _lowerCAmelCase = 'Hello world' _lowerCAmelCase = tokenizer(UpperCAmelCase_ , return_tensors='np' ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(UpperCAmelCase_ , 'do_samples' ): model.generate(UpperCAmelCase_ , do_samples=UpperCAmelCase_ ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(UpperCAmelCase_ , 'foo' ): _lowerCAmelCase = {'foo': 'bar'} model.generate(UpperCAmelCase_ , **UpperCAmelCase_ )
491
"""simple docstring""" def __snake_case ( SCREAMING_SNAKE_CASE: str ): """simple docstring""" _lowerCAmelCase = [0 for i in range(len(SCREAMING_SNAKE_CASE ) )] # initialize interval's left pointer and right pointer _lowerCAmelCase , _lowerCAmelCase = 0, 0 for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): # case when current index is inside the interval if i <= right_pointer: _lowerCAmelCase = min(right_pointer - i + 1 , z_result[i - left_pointer] ) _lowerCAmelCase = min_edge while go_next(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: _lowerCAmelCase , _lowerCAmelCase = i, i + z_result[i] - 1 return z_result def __snake_case ( SCREAMING_SNAKE_CASE: int , SCREAMING_SNAKE_CASE: list[int] , SCREAMING_SNAKE_CASE: str ): """simple docstring""" return i + z_result[i] < len(SCREAMING_SNAKE_CASE ) and s[z_result[i]] == s[i + z_result[i]] def __snake_case ( SCREAMING_SNAKE_CASE: str , SCREAMING_SNAKE_CASE: str ): """simple docstring""" _lowerCAmelCase = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string _lowerCAmelCase = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(SCREAMING_SNAKE_CASE ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
491
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Tuple = { 'configuration_graphormer': ['GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GraphormerConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ 'GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'GraphormerForGraphClassification', 'GraphormerModel', 'GraphormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
139
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 A_ : Optional[Any] = data_utils.TransfoXLTokenizer A_ : Union[str, Any] = data_utils.TransfoXLCorpus A_ : Any = data_utils A_ : Optional[Any] = data_utils def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) -> List[str]: if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCAmelCase__ , 'rb' ) as fp: UpperCamelCase_: Union[str, Any] = pickle.load(UpperCAmelCase__ , encoding='latin1' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) UpperCamelCase_: Any = pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['pretrained_vocab_file'] print(F'''Save vocabulary to {pytorch_vocab_dump_path}''' ) UpperCamelCase_: Union[str, Any] = corpus.vocab.__dict__ torch.save(UpperCAmelCase__ , UpperCAmelCase__ ) UpperCamelCase_: str = corpus.__dict__ corpus_dict_no_vocab.pop('vocab' , UpperCAmelCase__ ) UpperCamelCase_: str = pytorch_dump_folder_path + '/' + CORPUS_NAME print(F'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(UpperCAmelCase__ , UpperCAmelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model UpperCamelCase_: Any = os.path.abspath(UpperCAmelCase__ ) UpperCamelCase_: Dict = os.path.abspath(UpperCAmelCase__ ) print(F'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": UpperCamelCase_: List[str] = TransfoXLConfig() else: UpperCamelCase_: Optional[int] = TransfoXLConfig.from_json_file(UpperCAmelCase__ ) print(F'''Building PyTorch model from configuration: {config}''' ) UpperCamelCase_: Union[str, Any] = TransfoXLLMHeadModel(UpperCAmelCase__ ) UpperCamelCase_: Tuple = load_tf_weights_in_transfo_xl(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # Save pytorch-model UpperCamelCase_: str = os.path.join(UpperCAmelCase__ , UpperCAmelCase__ ) UpperCamelCase_: Union[str, Any] = os.path.join(UpperCAmelCase__ , UpperCAmelCase__ ) print(F'''Save PyTorch model to {os.path.abspath(UpperCAmelCase__ )}''' ) torch.save(model.state_dict() , UpperCAmelCase__ ) print(F'''Save configuration file to {os.path.abspath(UpperCAmelCase__ )}''' ) with open(UpperCAmelCase__ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A_ : int = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) A_ : Tuple = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
57
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCAmelCase : str = { """configuration_transfo_xl""": ["""TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TransfoXLConfig"""], """tokenization_transfo_xl""": ["""TransfoXLCorpus""", """TransfoXLTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ """TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST""", """AdaptiveEmbedding""", """TransfoXLForSequenceClassification""", """TransfoXLLMHeadModel""", """TransfoXLModel""", """TransfoXLPreTrainedModel""", """load_tf_weights_in_transfo_xl""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ """TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFAdaptiveEmbedding""", """TFTransfoXLForSequenceClassification""", """TFTransfoXLLMHeadModel""", """TFTransfoXLMainLayer""", """TFTransfoXLModel""", """TFTransfoXLPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys _UpperCAmelCase : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
188
from ..utils import DummyObject, requires_backends class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch""", """transformers""", """onnx"""] def __init__( self : Tuple , *UpperCAmelCase : str , **UpperCAmelCase : Optional[int] ) -> List[str]: requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : List[str] ) -> int: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Tuple ) -> Union[str, Any]: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch""", """transformers""", """onnx"""] def __init__( self : List[Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Dict ) -> Any: requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : str , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Union[str, Any] ) -> Tuple: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : str , **UpperCAmelCase : Dict ) -> str: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch""", """transformers""", """onnx"""] def __init__( self : Dict , *UpperCAmelCase : Dict , **UpperCAmelCase : Dict ) -> Optional[int]: requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Any ) -> Any: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Tuple: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch""", """transformers""", """onnx"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Optional[int] ) -> Any: requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : int ) -> Dict: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : Dict , **UpperCAmelCase : Optional[int] ) -> Optional[int]: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch""", """transformers""", """onnx"""] def __init__( self : List[str] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : List[str] ) -> Optional[int]: requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[Any] ) -> str: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Any ) -> Any: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch""", """transformers""", """onnx"""] def __init__( self : List[str] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Optional[int] ) -> Optional[Any]: requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Optional[Any] ) -> Tuple: requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : List[str] , **UpperCAmelCase : List[str] ) -> Tuple: requires_backends(cls , ['torch', 'transformers', 'onnx'] )
188
1
from ..utils import DummyObject, requires_backends class lowerCAmelCase__ ( metaclass=_lowerCamelCase ): A_ : List[Any] = ['note_seq'] def __init__( self : int , *__UpperCamelCase : Optional[int] , **__UpperCamelCase : List[Any] ) -> Optional[Any]: requires_backends(self , ['note_seq'] ) @classmethod def __UpperCamelCase ( cls : Optional[int] , *__UpperCamelCase : str , **__UpperCamelCase : Tuple ) -> str: requires_backends(cls , ['note_seq'] ) @classmethod def __UpperCamelCase ( cls : Optional[int] , *__UpperCamelCase : Union[str, Any] , **__UpperCamelCase : List[str] ) -> str: requires_backends(cls , ['note_seq'] )
106
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase__ ( _lowerCamelCase , unittest.TestCase ): A_ : Dict = GPTSanJapaneseTokenizer A_ : Optional[int] = False A_ : List[str] = {'do_clean_text': False, 'add_prefix_space': False} def __UpperCamelCase ( self : Tuple ) -> Any: super().setUp() # fmt: off A = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on A = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 A = {'unk_token': '<unk>'} A = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) A = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(__UpperCamelCase ) ) def __UpperCamelCase ( self : Any , **__UpperCamelCase : Any ) -> Tuple: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **__UpperCamelCase ) def __UpperCamelCase ( self : str , __UpperCamelCase : Dict ) -> List[str]: A = 'こんにちは、世界。 \nこんばんは、㔺界。😀' A = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def __UpperCamelCase ( self : str , __UpperCamelCase : Optional[Any] ) -> List[str]: A , A = self.get_input_output_texts(__UpperCamelCase ) A = tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) A = tokenizer.decode(__UpperCamelCase , clean_up_tokenization_spaces=__UpperCamelCase ) return text, ids def __UpperCamelCase ( self : Any ) -> int: pass # TODO add if relevant def __UpperCamelCase ( self : List[str] ) -> Tuple: pass # TODO add if relevant def __UpperCamelCase ( self : List[str] ) -> List[Any]: pass # TODO add if relevant def __UpperCamelCase ( self : List[str] ) -> int: A = self.get_tokenizer() # Testing tokenization A = 'こんにちは、世界。 こんばんは、㔺界。' A = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] A = tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) # Testing conversion to ids without special tokens A = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] A = tokenizer.convert_tokens_to_ids(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) # Testing conversion to ids with special tokens A = tokens + [tokenizer.unk_token] A = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] A = tokenizer.convert_tokens_to_ids(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) def __UpperCamelCase ( self : Union[str, Any] ) -> List[str]: A = self.get_tokenizer() # Testing tokenization A = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' A = 'こんにちは、、、、世界。こんばんは、、、、世界。' A = tokenizer.encode(__UpperCamelCase ) A = tokenizer.decode(__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) @slow def __UpperCamelCase ( self : Union[str, Any] ) -> str: A = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization A = 'こんにちは、世界。' A = 'こんばんは、㔺界。😀' A = 'こんにちは、世界。こんばんは、世界。😀' A = tokenizer.encode(prefix_text + input_text ) A = tokenizer.encode('' , prefix_text=prefix_text + input_text ) A = tokenizer.encode(__UpperCamelCase , prefix_text=__UpperCamelCase ) A = tokenizer.decode(__UpperCamelCase ) A = tokenizer.decode(__UpperCamelCase ) A = tokenizer.decode(__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) @slow def __UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: A = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization A = 'こんにちは、世界。' A = 'こんばんは、㔺界。😀' A = len(tokenizer.encode(__UpperCamelCase ) ) - 2 A = len(tokenizer.encode(__UpperCamelCase ) ) - 2 A = [1] + [0] * (len_prefix + len_text + 1) A = [1] * (len_prefix + len_text + 1) + [0] A = [1] + [1] * (len_prefix) + [0] * (len_text + 1) A = tokenizer(prefix_text + input_text ).token_type_ids A = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids A = tokenizer(__UpperCamelCase , prefix_text=__UpperCamelCase ).token_type_ids self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) @slow def __UpperCamelCase ( self : Any ) -> str: A = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) A = tokenizer.encode('あンいワ' ) A = tokenizer.encode('' , prefix_text='あンいワ' ) A = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(__UpperCamelCase ) , tokenizer.decode(__UpperCamelCase ) ) self.assertEqual(tokenizer.decode(__UpperCamelCase ) , tokenizer.decode(__UpperCamelCase ) ) self.assertNotEqual(__UpperCamelCase , __UpperCamelCase ) self.assertNotEqual(__UpperCamelCase , __UpperCamelCase ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def __UpperCamelCase ( self : Any ) -> List[Any]: A = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) A = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] A = tokenizer(__UpperCamelCase , padding=__UpperCamelCase ) A = tokenizer.batch_encode_plus(__UpperCamelCase , padding=__UpperCamelCase ) # fmt: off A = [[35_993, 8_640, 25_948, 35_998, 30_647, 35_675, 35_999, 35_999], [35_993, 10_382, 9_868, 35_998, 30_646, 9_459, 30_646, 35_675]] A = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] A = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , __UpperCamelCase ) self.assertListEqual(x_token.token_type_ids , __UpperCamelCase ) self.assertListEqual(x_token.attention_mask , __UpperCamelCase ) self.assertListEqual(x_token_a.input_ids , __UpperCamelCase ) self.assertListEqual(x_token_a.token_type_ids , __UpperCamelCase ) self.assertListEqual(x_token_a.attention_mask , __UpperCamelCase ) def __UpperCamelCase ( self : Tuple ) -> str: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def __UpperCamelCase ( self : Dict ) -> int: # tokenizer has no padding token pass
106
1
'''simple docstring''' import os from math import logaa def SCREAMING_SNAKE_CASE_ (UpperCamelCase = "base_exp.txt" ) -> int: lowerCamelCase__ : str = 0 lowerCamelCase__ : str = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(lowerCamelCase__ ) , lowerCamelCase__ ) ) ): lowerCamelCase__ , lowerCamelCase__ : Any = list(map(lowerCamelCase__ , line.split(""",""" ) ) ) if x * logaa(lowerCamelCase__ ) > largest: lowerCamelCase__ : Dict = x * logaa(lowerCamelCase__ ) lowerCamelCase__ : Dict = i + 1 return result if __name__ == "__main__": print(solution())
721
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _lowercase ( _lowercase ): def __init__( self: Optional[Any] , UpperCamelCase__: Any , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : Optional[int] = dataset lowerCamelCase__ : Optional[int] = process lowerCamelCase__ : List[str] = params def __len__( self: List[str] ): return len(self.dataset ) def __getitem__( self: Any , UpperCamelCase__: int ): lowerCamelCase__ : Dict = self.dataset[i] lowerCamelCase__ : Union[str, Any] = self.process(UpperCamelCase__ , **self.params ) return processed class _lowercase ( _lowercase ): def __init__( self: Optional[Any] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Tuple , UpperCamelCase__: Any=None ): lowerCamelCase__ : int = loader lowerCamelCase__ : str = infer lowerCamelCase__ : Optional[int] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : int = loader_batch_size # Internal bookkeeping lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Optional[Any] = None def __len__( self: Dict ): return len(self.loader ) def __iter__( self: Optional[int] ): lowerCamelCase__ : List[Any] = iter(self.loader ) return self def lowerCamelCase_ ( self: Any ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice lowerCamelCase__ : str = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) lowerCamelCase__ : int = {} for k, element in self._loader_batch_data.items(): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): # Convert ModelOutput to tuple first lowerCamelCase__ : str = element.to_tuple() if isinstance(element[0] , torch.Tensor ): lowerCamelCase__ : Tuple = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): lowerCamelCase__ : str = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(UpperCamelCase__ , UpperCamelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): lowerCamelCase__ : List[str] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): lowerCamelCase__ : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around lowerCamelCase__ : List[str] = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers lowerCamelCase__ : Optional[Any] = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers lowerCamelCase__ : int = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. lowerCamelCase__ : str = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 lowerCamelCase__ : Optional[int] = self._loader_batch_data.__class__(UpperCamelCase__ ) self._loader_batch_index += 1 return result def lowerCamelCase_ ( self: List[Any] ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch lowerCamelCase__ : Optional[Any] = next(self.iterator ) lowerCamelCase__ : List[str] = self.infer(UpperCamelCase__ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(UpperCamelCase__ , torch.Tensor ): lowerCamelCase__ : Optional[Any] = processed else: lowerCamelCase__ : Union[str, Any] = list(processed.keys() )[0] lowerCamelCase__ : Any = processed[key] if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : Any = len(UpperCamelCase__ ) else: lowerCamelCase__ : List[str] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. lowerCamelCase__ : Union[str, Any] = observed_batch_size # Setting internal index to unwrap the batch lowerCamelCase__ : List[Any] = processed lowerCamelCase__ : List[Any] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class _lowercase ( _lowercase ): def __init__( self: List[str] , UpperCamelCase__: Any , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any]=None ): super().__init__(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def __iter__( self: Union[str, Any] ): lowerCamelCase__ : str = iter(self.loader ) lowerCamelCase__ : int = None return self def lowerCamelCase_ ( self: str ): if self.subiterator is None: lowerCamelCase__ : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item lowerCamelCase__ : Tuple = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators lowerCamelCase__ : Any = self.infer(next(self.iterator ) , **self.params ) lowerCamelCase__ : Union[str, Any] = next(self.subiterator ) return processed class _lowercase ( _lowercase ): def __iter__( self: List[Any] ): lowerCamelCase__ : int = iter(self.loader ) return self def lowerCamelCase_ ( self: Tuple ): # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. lowerCamelCase__ : List[str] = False lowerCamelCase__ : Union[str, Any] = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: lowerCamelCase__ : Any = self.loader_batch_item() lowerCamelCase__ : Tuple = item.pop("""is_last""" ) accumulator.append(UpperCamelCase__ ) if is_last: return accumulator while not is_last: lowerCamelCase__ : str = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(UpperCamelCase__ , torch.Tensor ): lowerCamelCase__ : Dict = processed else: lowerCamelCase__ : Dict = list(processed.keys() )[0] lowerCamelCase__ : Dict = processed[key] if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : List[Any] = len(UpperCamelCase__ ) else: lowerCamelCase__ : Dict = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. lowerCamelCase__ : str = observed_batch_size lowerCamelCase__ : str = processed lowerCamelCase__ : Optional[int] = 0 while self._loader_batch_index < self.loader_batch_size: lowerCamelCase__ : List[Any] = self.loader_batch_item() lowerCamelCase__ : str = item.pop("""is_last""" ) accumulator.append(UpperCamelCase__ ) if is_last: return accumulator else: lowerCamelCase__ : Optional[Any] = processed lowerCamelCase__ : Optional[int] = item.pop("""is_last""" ) accumulator.append(UpperCamelCase__ ) return accumulator class _lowercase ( _lowercase ): def __init__( self: Optional[int] , UpperCamelCase__: Dataset , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = dataset lowerCamelCase__ : str = key def __len__( self: Optional[Any] ): return len(self.dataset ) def __getitem__( self: List[str] , UpperCamelCase__: Any ): return self.dataset[i][self.key] class _lowercase ( _lowercase ): def __init__( self: Optional[int] , UpperCamelCase__: Dataset , UpperCamelCase__: str , UpperCamelCase__: str ): lowerCamelCase__ : str = dataset lowerCamelCase__ : Dict = keya lowerCamelCase__ : List[str] = keya def __len__( self: str ): return len(self.dataset ) def __getitem__( self: List[str] , UpperCamelCase__: Union[str, Any] ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
631
0
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("1.0.0a"): raise Exception("requires fairseq >= 1.0.0a") logging.set_verbosity_info() _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = "Hello world! cécé herlolip" def _lowerCAmelCase ( lowercase : str , lowercase : str , lowercase : bool ) ->List[str]: """simple docstring""" lowercase__ = FairseqRobertaModel.from_pretrained(__snake_case ) roberta.eval() # disable dropout lowercase__ = roberta.model.encoder.sentence_encoder lowercase__ = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=5_1_4 , type_vocab_size=1 , layer_norm_eps=1e-5 , ) if classification_head: lowercase__ = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print('''Our RoBERTa config:''' , __snake_case ) lowercase__ = XLMRobertaXLForSequenceClassification(__snake_case ) if classification_head else XLMRobertaXLForMaskedLM(__snake_case ) model.eval() # Now let's copy all the weights. # Embeddings lowercase__ = roberta_sent_encoder.embed_tokens.weight lowercase__ = roberta_sent_encoder.embed_positions.weight lowercase__ = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. lowercase__ = roberta_sent_encoder.layer_norm.weight lowercase__ = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer lowercase__ = model.roberta.encoder.layer[i] lowercase__ = roberta_sent_encoder.layers[i] lowercase__ = layer.attention lowercase__ = roberta_layer.self_attn_layer_norm.weight lowercase__ = roberta_layer.self_attn_layer_norm.bias # self attention lowercase__ = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) lowercase__ = roberta_layer.self_attn.q_proj.weight lowercase__ = roberta_layer.self_attn.q_proj.bias lowercase__ = roberta_layer.self_attn.k_proj.weight lowercase__ = roberta_layer.self_attn.k_proj.bias lowercase__ = roberta_layer.self_attn.v_proj.weight lowercase__ = roberta_layer.self_attn.v_proj.bias # self-attention output lowercase__ = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape lowercase__ = roberta_layer.self_attn.out_proj.weight lowercase__ = roberta_layer.self_attn.out_proj.bias # this one is final layer norm lowercase__ = roberta_layer.final_layer_norm.weight lowercase__ = roberta_layer.final_layer_norm.bias # intermediate lowercase__ = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape lowercase__ = roberta_layer.fca.weight lowercase__ = roberta_layer.fca.bias # output lowercase__ = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape lowercase__ = roberta_layer.fca.weight lowercase__ = roberta_layer.fca.bias # end of layer if classification_head: lowercase__ = roberta.model.classification_heads["mnli"].dense.weight lowercase__ = roberta.model.classification_heads["mnli"].dense.bias lowercase__ = roberta.model.classification_heads["mnli"].out_proj.weight lowercase__ = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head lowercase__ = roberta.model.encoder.lm_head.dense.weight lowercase__ = roberta.model.encoder.lm_head.dense.bias lowercase__ = roberta.model.encoder.lm_head.layer_norm.weight lowercase__ = roberta.model.encoder.lm_head.layer_norm.bias lowercase__ = roberta.model.encoder.lm_head.weight lowercase__ = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. lowercase__ = roberta.encode(__snake_case ).unsqueeze(0 ) # batch of size 1 lowercase__ = model(__snake_case )[0] if classification_head: lowercase__ = roberta.model.classification_heads["mnli"](roberta.extract_features(__snake_case ) ) else: lowercase__ = roberta.model(__snake_case )[0] print(our_output.shape , their_output.shape ) lowercase__ = torch.max(torch.abs(our_output - their_output ) ).item() print(F'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 lowercase__ = torch.allclose(__snake_case , __snake_case , atol=1e-3 ) print('''Do both models output the same tensors?''' , '''🔥''' if success else '''💩''' ) if not success: raise Exception('''Something went wRoNg''' ) pathlib.Path(__snake_case ).mkdir(parents=__snake_case , exist_ok=__snake_case ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) if __name__ == "__main__": _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--roberta_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--classification_head", action="store_true", help="Whether to convert a final classification head." ) _lowerCAmelCase = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
161
from collections import defaultdict from math import gcd def lowercase_ ( __snake_case : int = 1_50_00_00 ) -> int: '''simple docstring''' snake_case__ :defaultdict = defaultdict(__snake_case ) snake_case__ :List[Any] = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __snake_case , 2 ): if gcd(__snake_case , __snake_case ) > 1: continue snake_case__ :Any = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__snake_case , limit + 1 , __snake_case ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
241
0
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __UpperCAmelCase : """simple docstring""" def __init__( self , __A , __A=13 , __A=3 , __A=True , __A=True , __A=0.1 , __A=0.1 , __A=224 , __A=1000 , __A=[3, 3, 6, 4] , __A=[48, 56, 112, 220] , ): __a = parent __a = batch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = num_labels __a = image_size __a = layer_depths __a = embed_dims def snake_case_ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.num_labels ) __a = self.get_config() return config, pixel_values, labels def snake_case_ ( self ): return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="""gelu""" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=__lowerCamelCase , layer_scale_init_value=1E-5 , ) def snake_case_ ( self , __A , __A , __A ): __a = SwiftFormerModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __a = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def snake_case_ ( self , __A , __A , __A ): __a = self.num_labels __a = SwiftFormerForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __a = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) __a = SwiftFormerForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case_ ( self ): (__a) = self.prepare_config_and_inputs() __a = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCAmelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () _lowerCamelCase = ( {"""feature-extraction""": SwiftFormerModel, """image-classification""": SwiftFormerForImageClassification} if is_torch_available() else {} ) _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def snake_case_ ( self ): __a = SwiftFormerModelTester(self ) __a = ConfigTester( self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def snake_case_ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""SwiftFormer does not use inputs_embeds""" ) def snake_case_ ( self ): pass def snake_case_ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(__lowerCamelCase ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , nn.Linear ) ) def snake_case_ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(__lowerCamelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def snake_case_ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def snake_case_ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def snake_case_ ( self ): for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = SwiftFormerModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @unittest.skip(reason="""SwiftFormer does not output attentions""" ) def snake_case_ ( self ): pass def snake_case_ ( self ): def check_hidden_states_output(__A , __A , __A ): __a = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): __a = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) __a = outputs.hidden_states __a = 8 self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(__lowerCamelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def snake_case_ ( self ): def _config_zero_init(__A ): __a = copy.deepcopy(__lowerCamelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(__lowerCamelCase , __lowerCamelCase , 1E-10 ) if isinstance(getattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , __lowerCamelCase ): __a = _config_zero_init(getattr(__lowerCamelCase , __lowerCamelCase ) ) setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return configs_no_init __a = self.model_tester.prepare_config_and_inputs_for_common() __a = _config_zero_init(__lowerCamelCase ) for model_class in self.all_model_classes: __a = model_class(config=__lowerCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def snake_case_ ( self ): pass def a (): __a = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case_ ( self ): return ViTImageProcessor.from_pretrained("""MBZUAI/swiftformer-xs""" ) if is_vision_available() else None @slow def snake_case_ ( self ): __a = SwiftFormerForImageClassification.from_pretrained("""MBZUAI/swiftformer-xs""" ).to(__lowerCamelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=__lowerCamelCase , return_tensors="""pt""" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(**__lowerCamelCase ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) __a = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1E-4 ) )
702
from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent SCREAMING_SNAKE_CASE = {'UserAgent': UserAgent().random} def a (lowerCAmelCase__ ): __a = script.contents[0] __a = json.loads(data[data.find("""{\"config\"""" ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class __UpperCAmelCase : """simple docstring""" def __init__( self , __A ): __a = f'''https://www.instagram.com/{username}/''' __a = self.get_json() def snake_case_ ( self ): __a = requests.get(self.url , headers=__A ).text __a = BeautifulSoup(__A , """html.parser""" ).find_all("""script""" ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self ): return f'''{self.__class__.__name__}(\'{self.username}\')''' def __str__( self ): return f'''{self.fullname} ({self.username}) is {self.biography}''' @property def snake_case_ ( self ): return self.user_data["username"] @property def snake_case_ ( self ): return self.user_data["full_name"] @property def snake_case_ ( self ): return self.user_data["biography"] @property def snake_case_ ( self ): return self.user_data["business_email"] @property def snake_case_ ( self ): return self.user_data["external_url"] @property def snake_case_ ( self ): return self.user_data["edge_followed_by"]["count"] @property def snake_case_ ( self ): return self.user_data["edge_follow"]["count"] @property def snake_case_ ( self ): return self.user_data["edge_owner_to_timeline_media"]["count"] @property def snake_case_ ( self ): return self.user_data["profile_pic_url_hd"] @property def snake_case_ ( self ): return self.user_data["is_verified"] @property def snake_case_ ( self ): return self.user_data["is_private"] def a (lowerCAmelCase__ = "github" ): import os if os.environ.get("""CI""" ): return # test failing on GitHub Actions __a = InstagramUser(lowerCAmelCase__ ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , lowerCAmelCase__ ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 150 assert instagram_user.number_of_followers > 120_000 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith("""https://instagram.""" ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE = InstagramUser('github') print(instagram_user) print(f'''{instagram_user.number_of_posts = }''') print(f'''{instagram_user.number_of_followers = }''') print(f'''{instagram_user.number_of_followings = }''') print(f'''{instagram_user.email = }''') print(f'''{instagram_user.website = }''') print(f'''{instagram_user.profile_picture_url = }''') print(f'''{instagram_user.is_verified = }''') print(f'''{instagram_user.is_private = }''')
209
0
"""simple docstring""" from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __lt__( self , _lowercase ) -> List[str]: '''simple docstring''' return self[-1] < other[-1] def __eq__( self , _lowercase ) -> Any: '''simple docstring''' return self[-1] == other[-1] def __lowerCAmelCase ( __UpperCamelCase : list ): '''simple docstring''' snake_case_ : list[Stack] = [] # sort into stacks for element in collection: snake_case_ : Any = Stack([element] ) snake_case_ : Any = bisect_left(__UpperCamelCase , __UpperCamelCase ) if i != len(__UpperCamelCase ): stacks[i].append(__UpperCamelCase ) else: stacks.append(__UpperCamelCase ) # use a heap-based merge to merge stack efficiently snake_case_ : List[str] = merge(*(reversed(__UpperCamelCase ) for stack in stacks) ) return collection if __name__ == "__main__": __lowerCAmelCase : List[str] = input('''Enter numbers separated by a comma:\n''').strip() __lowerCAmelCase : int = [int(item) for item in user_input.split(''',''')] print(patience_sort(unsorted))
58
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''roformer''' def __init__( self , _lowercase=5_0_0_0_0 , _lowercase=None , _lowercase=7_6_8 , _lowercase=1_2 , _lowercase=1_2 , _lowercase=3_0_7_2 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=1_5_3_6 , _lowercase=2 , _lowercase=0.02 , _lowercase=1E-12 , _lowercase=0 , _lowercase=False , _lowercase=True , **_lowercase , ) -> int: '''simple docstring''' super().__init__(pad_token_id=_lowercase , **_lowercase ) snake_case_ : str = vocab_size snake_case_ : Any = hidden_size if embedding_size is None else embedding_size snake_case_ : List[str] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : Tuple = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : List[str] = rotary_value snake_case_ : str = use_cache class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ : Any = {0: """batch""", 1: """sequence"""} snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
58
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class _UpperCamelCase( SCREAMING_SNAKE_CASE ): __A: int = """openai/whisper-base""" __A: Optional[Any] = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) __A: Optional[Any] = """transcriber""" __A: Union[str, Any] = WhisperProcessor __A: int = WhisperForConditionalGeneration __A: List[str] = ["""audio"""] __A: List[str] = ["""text"""] def a__ ( self : List[Any] , _lowerCamelCase : Dict ): return self.pre_processor(_lowerCamelCase , return_tensors="pt" ).input_features def a__ ( self : Tuple , _lowerCamelCase : int ): return self.model.generate(inputs=_lowerCamelCase ) def a__ ( self : Tuple , _lowerCamelCase : Optional[Any] ): return self.pre_processor.batch_decode(_lowerCamelCase , skip_special_tokens=_lowerCamelCase )[0]
328
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase = logging.get_logger(__name__) set_seed(7_7_0) __lowerCamelCase = { 'c_attn': 'att_proj', 'c_proj': 'out_proj', 'c_fc': 'in_proj', 'transformer.': '', 'h.': 'layers.', 'ln_1': 'layernorm_1', 'ln_2': 'layernorm_2', 'ln_f': 'layernorm_final', 'wpe': 'position_embeds_layer', 'wte': 'input_embeds_layer', } __lowerCamelCase = { 'text_small': { 'repo_id': 'suno/bark', 'file_name': 'text.pt', }, 'coarse_small': { 'repo_id': 'suno/bark', 'file_name': 'coarse.pt', }, 'fine_small': { 'repo_id': 'suno/bark', 'file_name': 'fine.pt', }, 'text': { 'repo_id': 'suno/bark', 'file_name': 'text_2.pt', }, 'coarse': { 'repo_id': 'suno/bark', 'file_name': 'coarse_2.pt', }, 'fine': { 'repo_id': 'suno/bark', 'file_name': 'fine_2.pt', }, } __lowerCamelCase = os.path.dirname(os.path.abspath(__file__)) __lowerCamelCase = os.path.join(os.path.expanduser('~'), '.cache') __lowerCamelCase = os.path.join(os.getenv('XDG_CACHE_HOME', default_cache_dir), 'suno', 'bark_v0') def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> Tuple: """simple docstring""" _UpperCAmelCase : str = model_type if use_small: key += "_small" return os.path.join(_SCREAMING_SNAKE_CASE , REMOTE_MODEL_PATHS[key]["file_name"] ) def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" os.makedirs(_SCREAMING_SNAKE_CASE , exist_ok=_SCREAMING_SNAKE_CASE ) hf_hub_download(repo_id=_SCREAMING_SNAKE_CASE , filename=_SCREAMING_SNAKE_CASE , local_dir=_SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="text" ) -> Dict: """simple docstring""" if model_type == "text": _UpperCAmelCase : int = BarkSemanticModel _UpperCAmelCase : Dict = BarkSemanticConfig _UpperCAmelCase : Any = BarkSemanticGenerationConfig elif model_type == "coarse": _UpperCAmelCase : Optional[Any] = BarkCoarseModel _UpperCAmelCase : int = BarkCoarseConfig _UpperCAmelCase : Any = BarkCoarseGenerationConfig elif model_type == "fine": _UpperCAmelCase : Dict = BarkFineModel _UpperCAmelCase : Optional[Any] = BarkFineConfig _UpperCAmelCase : List[str] = BarkFineGenerationConfig else: raise NotImplementedError() _UpperCAmelCase : Union[str, Any] = F"""{model_type}_small""" if use_small else model_type _UpperCAmelCase : str = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(_SCREAMING_SNAKE_CASE ): logger.info(F"""{model_type} model not found, downloading into `{CACHE_DIR}`.""" ) _download(model_info["repo_id"] , model_info["file_name"] ) _UpperCAmelCase : int = torch.load(_SCREAMING_SNAKE_CASE , map_location=_SCREAMING_SNAKE_CASE ) # this is a hack _UpperCAmelCase : Union[str, Any] = checkpoint["model_args"] if "input_vocab_size" not in model_args: _UpperCAmelCase : List[Any] = model_args["vocab_size"] _UpperCAmelCase : Union[str, Any] = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments _UpperCAmelCase : str = model_args.pop("n_head" ) _UpperCAmelCase : Optional[int] = model_args.pop("n_embd" ) _UpperCAmelCase : Optional[Any] = model_args.pop("n_layer" ) _UpperCAmelCase : Tuple = ConfigClass(**checkpoint["model_args"] ) _UpperCAmelCase : List[Any] = ModelClass(config=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Optional[int] = GenerationConfigClass() _UpperCAmelCase : str = model_generation_config _UpperCAmelCase : Optional[int] = checkpoint["model"] # fixup checkpoint _UpperCAmelCase : Union[str, Any] = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(_SCREAMING_SNAKE_CASE ): # replace part of the key with corresponding layer name in HF implementation _UpperCAmelCase : List[Any] = k[len(_SCREAMING_SNAKE_CASE ) :] for old_layer_name in new_layer_name_dict: _UpperCAmelCase : Optional[Any] = new_k.replace(_SCREAMING_SNAKE_CASE , new_layer_name_dict[old_layer_name] ) _UpperCAmelCase : str = state_dict.pop(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Union[str, Any] = set(state_dict.keys() ) - set(model.state_dict().keys() ) _UpperCAmelCase : List[Any] = {k for k in extra_keys if not k.endswith(".attn.bias" )} _UpperCAmelCase : str = set(model.state_dict().keys() ) - set(state_dict.keys() ) _UpperCAmelCase : Optional[int] = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(_SCREAMING_SNAKE_CASE ) != 0: raise ValueError(F"""extra keys found: {extra_keys}""" ) if len(_SCREAMING_SNAKE_CASE ) != 0: raise ValueError(F"""missing keys: {missing_keys}""" ) model.load_state_dict(_SCREAMING_SNAKE_CASE , strict=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Optional[Any] = model.num_parameters(exclude_embeddings=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Optional[Any] = checkpoint["best_val_loss"].item() logger.info(F"""model loaded: {round(n_params/1E6 , 1 )}M params, {round(_SCREAMING_SNAKE_CASE , 3 )} loss""" ) model.eval() model.to(_SCREAMING_SNAKE_CASE ) del checkpoint, state_dict return model def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="text" ) -> Tuple: """simple docstring""" if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() _UpperCAmelCase : Optional[Any] = "cpu" # do conversion on cpu _UpperCAmelCase : List[str] = _get_ckpt_path(_SCREAMING_SNAKE_CASE , use_small=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : List[str] = _load_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , model_type=_SCREAMING_SNAKE_CASE , use_small=_SCREAMING_SNAKE_CASE ) # load bark initial model _UpperCAmelCase : Union[str, Any] = _bark_load_model(_SCREAMING_SNAKE_CASE , "cpu" , model_type=_SCREAMING_SNAKE_CASE , use_small=_SCREAMING_SNAKE_CASE ) if model_type == "text": _UpperCAmelCase : Tuple = bark_model["model"] if model.num_parameters(exclude_embeddings=_SCREAMING_SNAKE_CASE ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model _UpperCAmelCase : Dict = 5 _UpperCAmelCase : List[str] = 1_0 if model_type in ["text", "coarse"]: _UpperCAmelCase : str = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) _UpperCAmelCase : Dict = bark_model(_SCREAMING_SNAKE_CASE )[0] _UpperCAmelCase : int = model(_SCREAMING_SNAKE_CASE ) # take last logits _UpperCAmelCase : Tuple = output_new_model_total.logits[:, [-1], :] else: _UpperCAmelCase : List[Any] = 3 _UpperCAmelCase : Any = 8 _UpperCAmelCase : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) _UpperCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Union[str, Any] = bark_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Optional[Any] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1E-3: raise ValueError("initial and new outputs are not equal" ) Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> List[str]: """simple docstring""" _UpperCAmelCase : List[Any] = os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase : str = BarkSemanticConfig.from_pretrained(os.path.join(_SCREAMING_SNAKE_CASE , "config.json" ) ) _UpperCAmelCase : Tuple = BarkCoarseConfig.from_pretrained(os.path.join(_SCREAMING_SNAKE_CASE , "config.json" ) ) _UpperCAmelCase : str = BarkFineConfig.from_pretrained(os.path.join(_SCREAMING_SNAKE_CASE , "config.json" ) ) _UpperCAmelCase : Dict = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) _UpperCAmelCase : List[Any] = BarkSemanticModel.from_pretrained(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : int = BarkCoarseModel.from_pretrained(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : int = BarkFineModel.from_pretrained(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : Optional[int] = EncodecModel.from_pretrained("facebook/encodec_24khz" ) _UpperCAmelCase : Dict = BarkConfig.from_sub_model_configs( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase : List[str] = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) _UpperCAmelCase : Any = BarkModel(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase : int = semantic _UpperCAmelCase : Tuple = coarseAcoustic _UpperCAmelCase : str = fineAcoustic _UpperCAmelCase : str = codec _UpperCAmelCase : Optional[Any] = bark_generation_config Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) bark.save_pretrained(_SCREAMING_SNAKE_CASE , repo_id=_SCREAMING_SNAKE_CASE , push_to_hub=_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('model_type', type=str, help='text, coarse or fine.') parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--is_small', action='store_true', help='convert the small version instead of the large.') __lowerCamelCase = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
328
1
def __a ( A__ : int ): if not isinstance(A__ , A__ ): SCREAMING_SNAKE_CASE = F"Input value of [number={number}] must be an integer" raise TypeError(A__ ) if number < 0: return False SCREAMING_SNAKE_CASE = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
16
from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Optional[int] = logging.get_logger(__name__) __A : Optional[Any] = { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class _SCREAMING_SNAKE_CASE ( __snake_case ): '''simple docstring''' lowerCamelCase__ = "gpt_neox" def __init__( self : Optional[int] , __lowerCamelCase : List[str]=50432 , __lowerCamelCase : int=6144 , __lowerCamelCase : Optional[Any]=44 , __lowerCamelCase : Tuple=64 , __lowerCamelCase : Optional[int]=24576 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Any=0.25 , __lowerCamelCase : List[Any]=10000 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : int=0.0 , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[Any]=2048 , __lowerCamelCase : Tuple=0.02 , __lowerCamelCase : Tuple=1e-5 , __lowerCamelCase : Dict=True , __lowerCamelCase : int=0 , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : List[str]=False , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Optional[int]=None , **__lowerCamelCase : str , ): super().__init__(bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase ) SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = rotary_pct SCREAMING_SNAKE_CASE = rotary_emb_base SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = hidden_dropout SCREAMING_SNAKE_CASE = classifier_dropout SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = layer_norm_eps SCREAMING_SNAKE_CASE = use_cache SCREAMING_SNAKE_CASE = tie_word_embeddings SCREAMING_SNAKE_CASE = use_parallel_residual SCREAMING_SNAKE_CASE = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( "The hidden size is not divisble by the number of attention heads! Make sure to update them!" ) def _snake_case ( self : Union[str, Any] ): if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __lowerCamelCase ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " f"got {self.rope_scaling}" ) SCREAMING_SNAKE_CASE = self.rope_scaling.get("type" , __lowerCamelCase ) SCREAMING_SNAKE_CASE = self.rope_scaling.get("factor" , __lowerCamelCase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(__lowerCamelCase , __lowerCamelCase ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
16
1
"""simple docstring""" import random def a__ ( lowerCAmelCase : int ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = num - 1 UpperCAmelCase__ : Dict = 0 while s % 2 == 0: UpperCAmelCase__ : Tuple = s // 2 t += 1 for _ in range(5 ): UpperCAmelCase__ : Optional[int] = random.randrange(2 , num - 1 ) UpperCAmelCase__ : Optional[Any] = pow(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if v != 1: UpperCAmelCase__ : Union[str, Any] = 0 while v != (num - 1): if i == t - 1: return False else: UpperCAmelCase__ : Optional[int] = i + 1 UpperCAmelCase__ : Optional[int] = (v**2) % num return True def a__ ( lowerCAmelCase : int ): '''simple docstring''' if num < 2: return False UpperCAmelCase__ : List[str] = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 101, 103, 107, 109, 113, 127, 131, 137, 139, 149, 151, 157, 163, 167, 173, 179, 181, 191, 193, 197, 199, 211, 223, 227, 229, 233, 239, 241, 251, 257, 263, 269, 271, 277, 281, 283, 293, 307, 311, 313, 317, 331, 337, 347, 349, 353, 359, 367, 373, 379, 383, 389, 397, 401, 409, 419, 421, 431, 433, 439, 443, 449, 457, 461, 463, 467, 479, 487, 491, 499, 503, 509, 521, 523, 541, 547, 557, 563, 569, 571, 577, 587, 593, 599, 601, 607, 613, 617, 619, 631, 641, 643, 647, 653, 659, 661, 673, 677, 683, 691, 701, 709, 719, 727, 733, 739, 743, 751, 757, 761, 769, 773, 787, 797, 809, 811, 821, 823, 827, 829, 839, 853, 857, 859, 863, 877, 881, 883, 887, 907, 911, 919, 929, 937, 941, 947, 953, 967, 971, 977, 983, 991, 997, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(lowerCAmelCase ) def a__ ( lowerCAmelCase : int = 1024 ): '''simple docstring''' while True: UpperCAmelCase__ : Tuple = random.randrange(2 ** (keysize - 1) , 2 ** (keysize) ) if is_prime_low_num(lowerCAmelCase ): return num if __name__ == "__main__": A__ : int = generate_large_prime() print(("""Prime number:""", num)) print(("""is_prime_low_num:""", is_prime_low_num(num)))
660
"""simple docstring""" from manim import * class _lowercase ( lowerCAmelCase_ ): '''simple docstring''' def lowerCAmelCase__ ( self )-> Tuple: UpperCAmelCase__ : str = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase__ : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase__ : List[Any] = [mem.copy() for i in range(6 )] UpperCAmelCase__ : Union[str, Any] = [mem.copy() for i in range(6 )] UpperCAmelCase__ : List[Any] = VGroup(*__UpperCamelCase ).arrange(__UpperCamelCase , buff=0 ) UpperCAmelCase__ : int = VGroup(*__UpperCamelCase ).arrange(__UpperCamelCase , buff=0 ) UpperCAmelCase__ : Tuple = VGroup(__UpperCamelCase , __UpperCamelCase ).arrange(__UpperCamelCase , buff=0 ) UpperCAmelCase__ : Tuple = Text("CPU" , font_size=24 ) UpperCAmelCase__ : Any = Group(__UpperCamelCase , __UpperCamelCase ).arrange(__UpperCamelCase , buff=0.5 , aligned_edge=__UpperCamelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__UpperCamelCase ) UpperCAmelCase__ : List[Any] = [mem.copy() for i in range(4 )] UpperCAmelCase__ : List[str] = VGroup(*__UpperCamelCase ).arrange(__UpperCamelCase , buff=0 ) UpperCAmelCase__ : Union[str, Any] = Text("GPU" , font_size=24 ) UpperCAmelCase__ : Dict = Group(__UpperCamelCase , __UpperCamelCase ).arrange(__UpperCamelCase , buff=0.5 , aligned_edge=__UpperCamelCase ) gpu.move_to([-1, -1, 0] ) self.add(__UpperCamelCase ) UpperCAmelCase__ : Optional[int] = [mem.copy() for i in range(6 )] UpperCAmelCase__ : List[str] = VGroup(*__UpperCamelCase ).arrange(__UpperCamelCase , buff=0 ) UpperCAmelCase__ : Tuple = Text("Model" , font_size=24 ) UpperCAmelCase__ : Dict = Group(__UpperCamelCase , __UpperCamelCase ).arrange(__UpperCamelCase , buff=0.5 , aligned_edge=__UpperCamelCase ) model.move_to([3, -1.0, 0] ) self.add(__UpperCamelCase ) UpperCAmelCase__ : List[str] = [] for i, rect in enumerate(__UpperCamelCase ): rect.set_stroke(__UpperCamelCase ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCAmelCase__ : int = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__UpperCamelCase , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__UpperCamelCase ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__UpperCamelCase , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__UpperCamelCase , buff=0.0 ) self.add(__UpperCamelCase ) cpu_targs.append(__UpperCamelCase ) UpperCAmelCase__ : List[Any] = [mem.copy() for i in range(6 )] UpperCAmelCase__ : Any = VGroup(*__UpperCamelCase ).arrange(__UpperCamelCase , buff=0 ) UpperCAmelCase__ : Tuple = Text("Loaded Checkpoint" , font_size=24 ) UpperCAmelCase__ : Any = Group(__UpperCamelCase , __UpperCamelCase ).arrange(__UpperCamelCase , aligned_edge=__UpperCamelCase , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) UpperCAmelCase__ : Optional[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase__ : Any = MarkupText( F"<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__UpperCamelCase , __UpperCamelCase ) UpperCAmelCase__ : str = MarkupText( F"<span fgcolor='{BLUE}'>●</span> Checkpoint" , font_size=18 , ) blue_text.next_to(__UpperCamelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) UpperCAmelCase__ : Optional[Any] = MarkupText( F"Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCamelCase ) , Write(__UpperCamelCase ) ) self.play(Write(__UpperCamelCase , run_time=1 ) , Create(__UpperCamelCase , run_time=1 ) ) UpperCAmelCase__ : Union[str, Any] = [] UpperCAmelCase__ : List[str] = [] for i, rect in enumerate(__UpperCamelCase ): UpperCAmelCase__ : Optional[Any] = fill.copy().set_fill(__UpperCamelCase , opacity=0.7 ) target.move_to(__UpperCamelCase ) first_animations.append(GrowFromCenter(__UpperCamelCase , run_time=1 ) ) UpperCAmelCase__ : List[str] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__UpperCamelCase , run_time=1.5 ) ) self.play(*__UpperCamelCase ) self.play(*__UpperCamelCase ) self.wait()
660
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer _lowerCAmelCase : int = logging.get_logger(__name__) _lowerCAmelCase : List[str] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _lowerCAmelCase : str = { 'vocab_file': { 'google/realm-cc-news-pretrained-embedder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt' ), 'google/realm-cc-news-pretrained-encoder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt' ), 'google/realm-cc-news-pretrained-scorer': ( 'https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt' ), 'google/realm-cc-news-pretrained-openqa': ( 'https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt' ), 'google/realm-orqa-nq-openqa': 'https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt', 'google/realm-orqa-nq-reader': 'https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt', 'google/realm-orqa-wq-openqa': 'https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt', 'google/realm-orqa-wq-reader': 'https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt', }, 'tokenizer_file': { 'google/realm-cc-news-pretrained-embedder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont' ), 'google/realm-cc-news-pretrained-encoder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json' ), 'google/realm-cc-news-pretrained-scorer': ( 'https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json' ), 'google/realm-cc-news-pretrained-openqa': ( 'https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json' ), 'google/realm-orqa-nq-openqa': ( 'https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json' ), 'google/realm-orqa-nq-reader': ( 'https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json' ), 'google/realm-orqa-wq-openqa': ( 'https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json' ), 'google/realm-orqa-wq-reader': ( 'https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json' ), }, } _lowerCAmelCase : Tuple = { 'google/realm-cc-news-pretrained-embedder': 5_12, 'google/realm-cc-news-pretrained-encoder': 5_12, 'google/realm-cc-news-pretrained-scorer': 5_12, 'google/realm-cc-news-pretrained-openqa': 5_12, 'google/realm-orqa-nq-openqa': 5_12, 'google/realm-orqa-nq-reader': 5_12, 'google/realm-orqa-wq-openqa': 5_12, 'google/realm-orqa-wq-reader': 5_12, } _lowerCAmelCase : Tuple = { 'google/realm-cc-news-pretrained-embedder': {'do_lower_case': True}, 'google/realm-cc-news-pretrained-encoder': {'do_lower_case': True}, 'google/realm-cc-news-pretrained-scorer': {'do_lower_case': True}, 'google/realm-cc-news-pretrained-openqa': {'do_lower_case': True}, 'google/realm-orqa-nq-openqa': {'do_lower_case': True}, 'google/realm-orqa-nq-reader': {'do_lower_case': True}, 'google/realm-orqa-wq-openqa': {'do_lower_case': True}, 'google/realm-orqa-wq-reader': {'do_lower_case': True}, } class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : Union[str, Any] = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE : List[str] = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE : Dict = PRETRAINED_INIT_CONFIGURATION __SCREAMING_SNAKE_CASE : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE : Any = RealmTokenizer def __init__( self : int , A : Optional[Any]=None , A : int=None , A : int=True , A : Dict="[UNK]" , A : List[str]="[SEP]" , A : Dict="[PAD]" , A : Dict="[CLS]" , A : Optional[Any]="[MASK]" , A : Optional[Any]=True , A : Any=None , **A : int , ): super().__init__( __snake_case , tokenizer_file=__snake_case , do_lower_case=__snake_case , unk_token=__snake_case , sep_token=__snake_case , pad_token=__snake_case , cls_token=__snake_case , mask_token=__snake_case , tokenize_chinese_chars=__snake_case , strip_accents=__snake_case , **__snake_case , ) _UpperCAmelCase : Union[str, Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , __snake_case ) != do_lower_case or normalizer_state.get("strip_accents" , __snake_case ) != strip_accents or normalizer_state.get("handle_chinese_chars" , __snake_case ) != tokenize_chinese_chars ): _UpperCAmelCase : str = getattr(__snake_case , normalizer_state.pop("type" ) ) _UpperCAmelCase : int = do_lower_case _UpperCAmelCase : List[Any] = strip_accents _UpperCAmelCase : List[Any] = tokenize_chinese_chars _UpperCAmelCase : List[str] = normalizer_class(**__snake_case ) _UpperCAmelCase : Optional[Any] = do_lower_case def snake_case_ ( self : List[str] , A : int , **A : Optional[int] ): _UpperCAmelCase : Tuple = PaddingStrategy.MAX_LENGTH _UpperCAmelCase : Tuple = text _UpperCAmelCase : Any = kwargs.pop("text_pair" , __snake_case ) _UpperCAmelCase : int = kwargs.pop("return_tensors" , __snake_case ) _UpperCAmelCase : int = { "input_ids": [], "attention_mask": [], "token_type_ids": [], } for idx, candidate_text in enumerate(__snake_case ): if batch_text_pair is not None: _UpperCAmelCase : List[str] = batch_text_pair[idx] else: _UpperCAmelCase : str = None _UpperCAmelCase : str = super().__call__(__snake_case , __snake_case , return_tensors=__snake_case , **__snake_case ) _UpperCAmelCase : Dict = encoded_candidates.get("input_ids" ) _UpperCAmelCase : Optional[Any] = encoded_candidates.get("attention_mask" ) _UpperCAmelCase : Optional[int] = encoded_candidates.get("token_type_ids" ) if encoded_input_ids is not None: output_data["input_ids"].append(__snake_case ) if encoded_attention_mask is not None: output_data["attention_mask"].append(__snake_case ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(__snake_case ) _UpperCAmelCase : Tuple = {key: item for key, item in output_data.items() if len(__snake_case ) != 0} return BatchEncoding(__snake_case , tensor_type=__snake_case ) def snake_case_ ( self : Union[str, Any] , A : Optional[Any] , A : Any=None ): _UpperCAmelCase : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case_ ( self : List[str] , A : Optional[int] , A : Optional[int] = None ): _UpperCAmelCase : Any = [self.sep_token_id] _UpperCAmelCase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case_ ( self : Optional[Any] , A : Tuple , A : List[str] = None ): _UpperCAmelCase : Tuple = self._tokenizer.model.save(__snake_case , name=__snake_case ) return tuple(__snake_case )
289
import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('0.12.2'): raise Exception('requires fairseq >= 0.12.2') if version.parse(fairseq.__version__) > version.parse('2'): raise Exception('requires fairseq < v2') logging.set_verbosity_info() a__ : Any = logging.get_logger(__name__) a__ : Tuple = 'Hello, World!' a__ : Any = 'en_XX' def UpperCAmelCase_ ( _UpperCAmelCase :str , _UpperCAmelCase :str , _UpperCAmelCase :bool ) -> Optional[Any]: '''simple docstring''' A_ = Path('''data_bin''' ) A_ = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(_UpperCAmelCase ).parent ) , checkpoint_file=Path(_UpperCAmelCase ).name , _name='''xmod_base''' , arch='''xmod_base''' , task='''multilingual_masked_lm''' , data_name_or_path=str(_UpperCAmelCase ) , bpe='''sentencepiece''' , sentencepiece_model=str(Path(_UpperCAmelCase ).parent / '''sentencepiece.bpe.model''' ) , src_dict=str(data_dir / '''dict.txt''' ) , ) xmod.eval() # disable dropout print(_UpperCAmelCase ) A_ = xmod.model.encoder.sentence_encoder A_ = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings , hidden_size=xmod.cfg.model.encoder_embed_dim , num_hidden_layers=xmod.cfg.model.encoder_layers , num_attention_heads=xmod.cfg.model.encoder_attention_heads , intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1e-5 , pre_norm=xmod.cfg.model.encoder_normalize_before , adapter_reduction_factor=getattr(xmod.cfg.model , '''bottleneck''' , 2 ) , adapter_layer_norm=xmod.cfg.model.adapter_layer_norm , adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm , ln_before_adapter=xmod.cfg.model.ln_before_adapter , languages=xmod.cfg.model.languages , ) if classification_head: A_ = xmod.model.classification_heads['''mnli'''].out_proj.weight.shape[0] print('''Our X-MOD config:''' , _UpperCAmelCase ) A_ = XmodForSequenceClassification(_UpperCAmelCase ) if classification_head else XmodForMaskedLM(_UpperCAmelCase ) model.eval() # Now let's copy all the weights. # Embeddings A_ = xmod_sent_encoder.embed_tokens.weight A_ = xmod_sent_encoder.embed_positions.weight A_ = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. A_ = xmod_sent_encoder.layernorm_embedding.weight A_ = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer A_ = model.roberta.encoder.layer[i] A_ = xmod_sent_encoder.layers[i] # self attention A_ = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError('''Dimensions of self-attention weights do not match.''' ) A_ = xmod_layer.self_attn.q_proj.weight A_ = xmod_layer.self_attn.q_proj.bias A_ = xmod_layer.self_attn.k_proj.weight A_ = xmod_layer.self_attn.k_proj.bias A_ = xmod_layer.self_attn.v_proj.weight A_ = xmod_layer.self_attn.v_proj.bias # self-attention output A_ = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError('''Dimensions of self-attention output weights do not match.''' ) A_ = xmod_layer.self_attn.out_proj.weight A_ = xmod_layer.self_attn.out_proj.bias A_ = xmod_layer.self_attn_layer_norm.weight A_ = xmod_layer.self_attn_layer_norm.bias # intermediate A_ = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('''Dimensions of intermediate weights do not match.''' ) A_ = xmod_layer.fca.weight A_ = xmod_layer.fca.bias # output A_ = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('''Dimensions of feed-forward weights do not match.''' ) A_ = xmod_layer.fca.weight A_ = xmod_layer.fca.bias A_ = xmod_layer.final_layer_norm.weight A_ = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: A_ = xmod_layer.adapter_layer_norm.weight A_ = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError('''Lists of language adapters do not match.''' ) for lang_code, adapter in xmod_layer.adapter_modules.items(): A_ = bert_output.adapter_modules[lang_code] A_ = xmod_layer.adapter_modules[lang_code] A_ = from_adapter.fca.weight A_ = from_adapter.fca.bias A_ = from_adapter.fca.weight A_ = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: A_ = xmod_sent_encoder.layer_norm.weight A_ = xmod_sent_encoder.layer_norm.bias if classification_head: A_ = xmod.model.classification_heads['''mnli'''].dense.weight A_ = xmod.model.classification_heads['''mnli'''].dense.bias A_ = xmod.model.classification_heads['''mnli'''].out_proj.weight A_ = xmod.model.classification_heads['''mnli'''].out_proj.bias else: # LM Head A_ = xmod.model.encoder.lm_head.dense.weight A_ = xmod.model.encoder.lm_head.dense.bias A_ = xmod.model.encoder.lm_head.layer_norm.weight A_ = xmod.model.encoder.lm_head.layer_norm.bias A_ = xmod.model.encoder.lm_head.weight A_ = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. A_ = xmod.encode(_UpperCAmelCase ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(_UpperCAmelCase ) A_ = model(_UpperCAmelCase )[0] if classification_head: A_ = xmod.model.classification_heads['''mnli'''](xmod.extract_features(_UpperCAmelCase ) ) else: A_ = xmod.model(_UpperCAmelCase , lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape , their_output.shape ) A_ = torch.max(torch.abs(our_output - their_output ) ).item() print(f'max_absolute_diff = {max_absolute_diff}' ) # ~ 1e-7 A_ = torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1e-3 ) print('''Do both models output the same tensors?''' , '''🔥''' if success else '''💩''' ) if not success: raise Exception('''Something went wRoNg''' ) Path(_UpperCAmelCase ).mkdir(parents=_UpperCAmelCase , exist_ok=_UpperCAmelCase ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a__ : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xmod_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) a__ : Tuple = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
188
0
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device _snake_case : Union[str, Any] = False class lowerCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class lowerCAmelCase ( unittest.TestCase ): def lowercase ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self ): _SCREAMING_SNAKE_CASE = VersatileDiffusionTextToImagePipeline.from_pretrained("shi-labs/versatile-diffusion" ) # remove text_unet pipe.remove_unused_weights() pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) _SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger """ _SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe( prompt=_UpperCamelCase , generator=_UpperCamelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_UpperCamelCase ) _SCREAMING_SNAKE_CASE = VersatileDiffusionTextToImagePipeline.from_pretrained(_UpperCamelCase ) pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) _SCREAMING_SNAKE_CASE = generator.manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe( prompt=_UpperCamelCase , generator=_UpperCamelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def lowercase ( self ): _SCREAMING_SNAKE_CASE = VersatileDiffusionTextToImagePipeline.from_pretrained( "shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) _SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger """ _SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe( prompt=_UpperCamelCase , generator=_UpperCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images _SCREAMING_SNAKE_CASE = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE = np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
715
'''simple docstring''' import numpy # List of input, output pairs _snake_case : Optional[Any] = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) _snake_case : Any = (((515, 22, 13), 555), ((61, 35, 49), 150)) _snake_case : Any = [2, 4, 1, 5] _snake_case : str = len(train_data) _snake_case : int = 0.009 def _a ( _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : List[Any]="train" ): return calculate_hypothesis_value(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) - output( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _a ( _SCREAMING_SNAKE_CASE : List[Any] ): _SCREAMING_SNAKE_CASE = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def _a ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any] ): if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def _a ( _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : str ): if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def _a ( _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[int]=m ): _SCREAMING_SNAKE_CASE = 0 for i in range(_SCREAMING_SNAKE_CASE ): if index == -1: summation_value += _error(_SCREAMING_SNAKE_CASE ) else: summation_value += _error(_SCREAMING_SNAKE_CASE ) * train_data[i][0][index] return summation_value def _a ( _SCREAMING_SNAKE_CASE : Optional[Any] ): _SCREAMING_SNAKE_CASE = summation_of_cost_derivative(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) / m return cost_derivative_value def _a ( ): global parameter_vector # Tune these values to set a tolerance value for predicted output _SCREAMING_SNAKE_CASE = 0.000_002 _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 0 while True: j += 1 _SCREAMING_SNAKE_CASE = [0, 0, 0, 0] for i in range(0 , len(_SCREAMING_SNAKE_CASE ) ): _SCREAMING_SNAKE_CASE = get_cost_derivative(i - 1 ) _SCREAMING_SNAKE_CASE = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE , rtol=_SCREAMING_SNAKE_CASE , ): break _SCREAMING_SNAKE_CASE = temp_parameter_vector print(("Number of iterations:", j) ) def _a ( ): for i in range(len(_SCREAMING_SNAKE_CASE ) ): print(("Actual output value:", output(_SCREAMING_SNAKE_CASE , "test" )) ) print(("Hypothesis output:", calculate_hypothesis_value(_SCREAMING_SNAKE_CASE , "test" )) ) if __name__ == "__main__": run_gradient_descent() print("""\nTesting gradient descent for a linear hypothesis function.\n""") test_gradient_descent()
493
0
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch A = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[int] ,UpperCamelCase : str = None ,UpperCamelCase : uuid.UUID = None ,UpperCamelCase : Tuple=None ,UpperCamelCase : Union[str, Any]=None ) -> int: if not conversation_id: _lowercase : List[str] = uuid.uuida() if past_user_inputs is None: _lowercase : Dict = [] if generated_responses is None: _lowercase : List[Any] = [] _lowercase : uuid.UUID = conversation_id _lowercase : List[str] = past_user_inputs _lowercase : List[str] = generated_responses _lowercase : Optional[str] = text def __eq__( self : Tuple ,UpperCamelCase : Optional[int] ) -> Union[str, Any]: if not isinstance(UpperCamelCase ,UpperCamelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def _lowerCamelCase ( self : Dict ,UpperCamelCase : str ,UpperCamelCase : bool = False ) -> List[Any]: if self.new_user_input: if overwrite: logger.warning( F'''User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ''' F'''with: "{text}".''' ) _lowercase : Optional[Any] = text else: logger.warning( F'''User input added while unprocessed input was existing: "{self.new_user_input}" new input ''' F'''ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input''' ) else: _lowercase : Dict = text def _lowerCamelCase ( self : List[str] ) -> int: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _lowercase : Optional[int] = None def _lowerCamelCase ( self : List[str] ,UpperCamelCase : str ) -> Optional[Any]: self.generated_responses.append(UpperCamelCase ) def _lowerCamelCase ( self : Tuple ) -> List[Any]: for user_input, generated_response in zip(self.past_user_inputs ,self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Any ) -> str: _lowercase : Dict = F'''Conversation id: {self.uuid} \n''' for is_user, text in self.iter_texts(): _lowercase : str = 'user' if is_user else 'bot' output += F'''{name} >> {text} \n''' return output @add_end_docstrings( lowerCAmelCase__ , r"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): '''simple docstring''' def __init__( self : Dict ,*UpperCamelCase : Optional[Any] ,**UpperCamelCase : Tuple ) -> Tuple: super().__init__(*UpperCamelCase ,**UpperCamelCase ) if self.tokenizer.pad_token_id is None: _lowercase : Optional[Any] = self.tokenizer.eos_token def _lowerCamelCase ( self : List[str] ,UpperCamelCase : Tuple=None ,UpperCamelCase : Optional[int]=None ,UpperCamelCase : str=None ,**UpperCamelCase : Optional[int] ) -> Tuple: _lowercase : Dict = {} _lowercase : List[Any] = {} _lowercase : List[str] = {} if min_length_for_response is not None: _lowercase : Optional[int] = min_length_for_response if minimum_tokens is not None: _lowercase : Optional[Any] = minimum_tokens if "max_length" in generate_kwargs: _lowercase : Optional[Any] = generate_kwargs['max_length'] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _lowercase : str = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCamelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self : int ,UpperCamelCase : Union[Conversation, List[Conversation]] ,UpperCamelCase : Tuple=0 ,**UpperCamelCase : Optional[Any] ) -> Optional[int]: _lowercase : Optional[int] = super().__call__(UpperCamelCase ,num_workers=UpperCamelCase ,**UpperCamelCase ) if isinstance(UpperCamelCase ,UpperCamelCase ) and len(UpperCamelCase ) == 1: return outputs[0] return outputs def _lowerCamelCase ( self : Union[str, Any] ,UpperCamelCase : Conversation ,UpperCamelCase : List[Any]=32 ) -> Dict[str, Any]: if not isinstance(UpperCamelCase ,UpperCamelCase ): raise ValueError('ConversationalPipeline, expects Conversation as inputs' ) if conversation.new_user_input is None: raise ValueError( F'''Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ''' 'Add user inputs with the conversation\'s `add_user_input` method' ) if hasattr(self.tokenizer ,'_build_conversation_input_ids' ): _lowercase : Optional[int] = self.tokenizer._build_conversation_input_ids(UpperCamelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _lowercase : List[Any] = self._legacy_parse_and_tokenize(UpperCamelCase ) if self.framework == "pt": _lowercase : Tuple = torch.LongTensor([input_ids] ) elif self.framework == "tf": _lowercase : int = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def _lowerCamelCase ( self : Tuple ,UpperCamelCase : Any ,UpperCamelCase : List[str]=10 ,**UpperCamelCase : Dict ) -> int: _lowercase : Any = generate_kwargs.get('max_length' ,self.model.config.max_length ) _lowercase : str = model_inputs['input_ids'].shape[1] if max_length - minimum_tokens < n: logger.warning(F'''Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})''' ) _lowercase : int = max_length - minimum_tokens _lowercase : str = model_inputs['input_ids'][:, -trim:] if "attention_mask" in model_inputs: _lowercase : Union[str, Any] = model_inputs['attention_mask'][:, -trim:] _lowercase : Dict = model_inputs.pop('conversation' ) _lowercase : str = max_length _lowercase : Union[str, Any] = self.model.generate(**UpperCamelCase ,**UpperCamelCase ) if self.model.config.is_encoder_decoder: _lowercase : List[Any] = 1 else: _lowercase : List[str] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def _lowerCamelCase ( self : List[Any] ,UpperCamelCase : Tuple ,UpperCamelCase : Any=True ) -> int: _lowercase : Dict = model_outputs['output_ids'] _lowercase : Any = self.tokenizer.decode( output_ids[0] ,skip_special_tokens=UpperCamelCase ,clean_up_tokenization_spaces=UpperCamelCase ,) _lowercase : str = model_outputs['conversation'] conversation.mark_processed() conversation.append_response(UpperCamelCase ) return conversation def _lowerCamelCase ( self : Any ,UpperCamelCase : Conversation ) -> Dict: _lowercase : str = self.tokenizer.eos_token_id _lowercase : Any = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCamelCase ,add_special_tokens=UpperCamelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCamelCase ,add_special_tokens=UpperCamelCase ) ) if len(UpperCamelCase ) > self.tokenizer.model_max_length: _lowercase : Any = input_ids[-self.tokenizer.model_max_length :] return input_ids
125
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter A = '''Create a default config file for Accelerate with only a few flags set.''' def SCREAMING_SNAKE_CASE ( lowerCAmelCase__ : Any="no" , lowerCAmelCase__ : str = default_json_config_file , lowerCAmelCase__ : bool = False) -> Optional[int]: '''simple docstring''' _lowercase : List[Any] = Path(lowerCAmelCase__) path.parent.mkdir(parents=lowerCAmelCase__ , exist_ok=lowerCAmelCase__) if path.exists(): print( F'''Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.''') return False _lowercase : int = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( F'''`mixed_precision` should be one of \'no\', \'fp16\', \'bf16\', or \'fp8\'. Received {mixed_precision}''') _lowercase : str = { 'compute_environment': 'LOCAL_MACHINE', 'mixed_precision': mixed_precision, } if torch.cuda.is_available(): _lowercase : str = torch.cuda.device_count() _lowercase : Optional[Any] = num_gpus _lowercase : Tuple = False if num_gpus > 1: _lowercase : Tuple = 'MULTI_GPU' else: _lowercase : Any = 'NO' elif is_xpu_available() and use_xpu: _lowercase : List[str] = torch.xpu.device_count() _lowercase : Union[str, Any] = num_xpus _lowercase : List[Any] = False if num_xpus > 1: _lowercase : Tuple = 'MULTI_XPU' else: _lowercase : List[str] = 'NO' elif is_npu_available(): _lowercase : Optional[int] = torch.npu.device_count() _lowercase : Any = num_npus _lowercase : List[Any] = False if num_npus > 1: _lowercase : int = 'MULTI_NPU' else: _lowercase : Dict = 'NO' else: _lowercase : List[Any] = 0 _lowercase : Optional[int] = True _lowercase : List[str] = 1 _lowercase : List[Any] = 'NO' _lowercase : Union[str, Any] = ClusterConfig(**lowerCAmelCase__) config.to_json_file(lowerCAmelCase__) return path def SCREAMING_SNAKE_CASE ( lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : List[str]) -> Union[str, Any]: '''simple docstring''' _lowercase : Tuple = parser.add_parser('default' , parents=lowerCAmelCase__ , help=lowerCAmelCase__ , formatter_class=lowerCAmelCase__) parser.add_argument( '--config_file' , default=lowerCAmelCase__ , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , dest='save_location' , ) parser.add_argument( '--mixed_precision' , choices=['no', 'fp16', 'bf16'] , type=lowerCAmelCase__ , help='Whether or not to use mixed precision training. ' 'Choose between FP16 and BF16 (bfloat16) training. ' 'BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.' , default='no' , ) parser.set_defaults(func=lowerCAmelCase__) return parser def SCREAMING_SNAKE_CASE ( lowerCAmelCase__ : List[Any]) -> Dict: '''simple docstring''' _lowercase : Tuple = write_basic_config(args.mixed_precision , args.save_location) if config_file: print(F'''accelerate configuration saved at {config_file}''')
125
1
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): lowerCamelCase__ : Any =["speech"] def __init__( self , *lowerCamelCase , **lowerCamelCase ) -> Dict: """simple docstring""" requires_backends(self , ['''speech'''] ) class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): lowerCamelCase__ : Optional[int] =["speech"] def __init__( self , *lowerCamelCase , **lowerCamelCase ) -> Any: """simple docstring""" requires_backends(self , ['''speech'''] )
336
# Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar lowercase_ = TypeVar('''T''') class A__ ( Generic[T] ): def __init__( self , lowerCamelCase = True ) -> None: """simple docstring""" __magic_name__ : dict[T, list[T]] = {} # dictionary of lists __magic_name__ : str = directed def lowercase ( self , lowerCamelCase , lowerCamelCase ) -> GraphAdjacencyList[T]: """simple docstring""" if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCamelCase ) self.adj_list[destination_vertex].append(lowerCamelCase ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCamelCase ) __magic_name__ : Optional[Any] = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(lowerCamelCase ) __magic_name__ : Union[str, Any] = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: __magic_name__ : Optional[int] = [destination_vertex] __magic_name__ : Optional[Any] = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCamelCase ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCamelCase ) __magic_name__ : List[str] = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: __magic_name__ : Dict = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: __magic_name__ : List[str] = [destination_vertex] __magic_name__ : List[str] = [] return self def __repr__( self ) -> str: """simple docstring""" return pformat(self.adj_list )
336
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' __A : Union[List[np.ndarray], torch.FloatTensor] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
340
import copy import random from transformers import CLIPTokenizer class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' def __init__( self , *__A , **__A ): """simple docstring""" super().__init__(*__A , **__A ) lowerCamelCase : Dict = {} def _snake_case ( self , __A , *__A , **__A ): """simple docstring""" lowerCamelCase : int = super().add_tokens(__A , *__A , **__A ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" " `placeholder_token` that is not already in the tokenizer." ) def _snake_case ( self , __A , *__A , __A=1 , **__A ): """simple docstring""" lowerCamelCase : Optional[Any] = [] if num_vec_per_token == 1: self.try_adding_tokens(__A , *__A , **__A ) output.append(__A ) else: lowerCamelCase : Any = [] for i in range(__A ): lowerCamelCase : List[str] = placeholder_token + F"""_{i}""" self.try_adding_tokens(__A , *__A , **__A ) output.append(__A ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) lowerCamelCase : Tuple = output def _snake_case ( self , __A , __A=False , __A=1.0 ): """simple docstring""" if isinstance(__A , __A ): lowerCamelCase : Optional[Any] = [] for i in range(len(__A ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=__A ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: lowerCamelCase : Optional[int] = self.token_map[placeholder_token] lowerCamelCase : str = tokens[: 1 + int(len(__A ) * prop_tokens_to_load )] if vector_shuffle: lowerCamelCase : List[str] = copy.copy(__A ) random.shuffle(__A ) lowerCamelCase : Any = text.replace(__A , " ".join(__A ) ) return text def __call__( self , __A , *__A , __A=False , __A=1.0 , **__A ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( __A , vector_shuffle=__A , prop_tokens_to_load=__A ) , *__A , **__A , ) def _snake_case ( self , __A , *__A , __A=False , __A=1.0 , **__A ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( __A , vector_shuffle=__A , prop_tokens_to_load=__A ) , *__A , **__A , )
340
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class __snake_case ( lowerCAmelCase__ ): __lowerCAmelCase : Union[List[np.ndarray], torch.FloatTensor] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
620
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __snake_case ( lowerCAmelCase__ ): def __init__( self , _A , _A , _A , _A = None , ): super().__init__() self.register_modules(transformer=_A , vae=_A , scheduler=_A) # create a imagenet -> id dictionary for easier use SCREAMING_SNAKE_CASE_ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split(','): SCREAMING_SNAKE_CASE_ = int(_A) SCREAMING_SNAKE_CASE_ = dict(sorted(self.labels.items())) def lowerCAmelCase__ ( self , _A): if not isinstance(_A , _A): SCREAMING_SNAKE_CASE_ = list(_A) for l in label: if l not in self.labels: raise ValueError( f"""{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.""") return [self.labels[l] for l in label] @torch.no_grad() def __call__( self , _A , _A = 4.0 , _A = None , _A = 50 , _A = "pil" , _A = True , ): SCREAMING_SNAKE_CASE_ = len(_A) SCREAMING_SNAKE_CASE_ = self.transformer.config.sample_size SCREAMING_SNAKE_CASE_ = self.transformer.config.in_channels SCREAMING_SNAKE_CASE_ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_A , device=self.device , dtype=self.transformer.dtype , ) SCREAMING_SNAKE_CASE_ = torch.cat([latents] * 2) if guidance_scale > 1 else latents SCREAMING_SNAKE_CASE_ = torch.tensor(_A , device=self.device).reshape(-1) SCREAMING_SNAKE_CASE_ = torch.tensor([1000] * batch_size , device=self.device) SCREAMING_SNAKE_CASE_ = torch.cat([class_labels, class_null] , 0) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_A) for t in self.progress_bar(self.scheduler.timesteps): if guidance_scale > 1: SCREAMING_SNAKE_CASE_ = latent_model_input[: len(_A) // 2] SCREAMING_SNAKE_CASE_ = torch.cat([half, half] , dim=0) SCREAMING_SNAKE_CASE_ = self.scheduler.scale_model_input(_A , _A) SCREAMING_SNAKE_CASE_ = t if not torch.is_tensor(_A): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) SCREAMING_SNAKE_CASE_ = latent_model_input.device.type == 'mps' if isinstance(_A , _A): SCREAMING_SNAKE_CASE_ = torch.floataa if is_mps else torch.floataa else: SCREAMING_SNAKE_CASE_ = torch.intaa if is_mps else torch.intaa SCREAMING_SNAKE_CASE_ = torch.tensor([timesteps] , dtype=_A , device=latent_model_input.device) elif len(timesteps.shape) == 0: SCREAMING_SNAKE_CASE_ = timesteps[None].to(latent_model_input.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML SCREAMING_SNAKE_CASE_ = timesteps.expand(latent_model_input.shape[0]) # predict noise model_output SCREAMING_SNAKE_CASE_ = self.transformer( _A , timestep=_A , class_labels=_A).sample # perform guidance if guidance_scale > 1: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = torch.split(_A , len(_A) // 2 , dim=0) SCREAMING_SNAKE_CASE_ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) SCREAMING_SNAKE_CASE_ = torch.cat([half_eps, half_eps] , dim=0) SCREAMING_SNAKE_CASE_ = torch.cat([eps, rest] , dim=1) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = torch.split(_A , _A , dim=1) else: SCREAMING_SNAKE_CASE_ = noise_pred # compute previous image: x_t -> x_t-1 SCREAMING_SNAKE_CASE_ = self.scheduler.step(_A , _A , _A).prev_sample if guidance_scale > 1: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = latent_model_input.chunk(2 , dim=0) else: SCREAMING_SNAKE_CASE_ = latent_model_input SCREAMING_SNAKE_CASE_ = 1 / self.vae.config.scaling_factor * latents SCREAMING_SNAKE_CASE_ = self.vae.decode(_A).sample SCREAMING_SNAKE_CASE_ = (samples / 2 + 0.5).clamp(0 , 1) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 SCREAMING_SNAKE_CASE_ = samples.cpu().permute(0 , 2 , 3 , 1).float().numpy() if output_type == "pil": SCREAMING_SNAKE_CASE_ = self.numpy_to_pil(_A) if not return_dict: return (samples,) return ImagePipelineOutput(images=_A)
620
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class lowercase__ ( unittest.TestCase ): def __init__( self : Tuple , UpperCAmelCase_ : Any , UpperCAmelCase_ : int=7 , UpperCAmelCase_ : Optional[int]=3 , UpperCAmelCase_ : Tuple=18 , UpperCAmelCase_ : Union[str, Any]=30 , UpperCAmelCase_ : Union[str, Any]=400 , UpperCAmelCase_ : Union[str, Any]=True , UpperCAmelCase_ : Union[str, Any]=32 , UpperCAmelCase_ : List[Any]=True , ): SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = min_resolution SCREAMING_SNAKE_CASE__ = max_resolution SCREAMING_SNAKE_CASE__ = do_resize SCREAMING_SNAKE_CASE__ = size_divisor SCREAMING_SNAKE_CASE__ = do_rescale def A_ ( self : Dict ): return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class lowercase__ ( _UpperCAmelCase , unittest.TestCase ): A__ : Dict =GLPNImageProcessor if is_vision_available() else None def A_ ( self : Any ): SCREAMING_SNAKE_CASE__ = GLPNImageProcessingTester(self ) @property def A_ ( self : Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def A_ ( self : Dict ): SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase_ , 'do_resize' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , 'size_divisor' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , 'resample' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , 'do_rescale' ) ) def A_ ( self : Tuple ): pass def A_ ( self : int ): # Initialize image_processing SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) SCREAMING_SNAKE_CASE__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def A_ ( self : List[Any] ): # Initialize image_processing SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , numpify=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) SCREAMING_SNAKE_CASE__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def A_ ( self : Optional[int] ): # Initialize image_processing SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , torchify=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) SCREAMING_SNAKE_CASE__ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
472
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def _lowercase ( UpperCamelCase_ ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(UpperCamelCase_ , UpperCamelCase_ ) def _lowercase ( UpperCamelCase_ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = emb.weight.shape SCREAMING_SNAKE_CASE__ = nn.Linear(UpperCamelCase_ , UpperCamelCase_ , bias=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = emb.weight.data return lin_layer def _lowercase ( UpperCamelCase_ , UpperCamelCase_="facebook/mbart-large-en-ro" , UpperCamelCase_=False , UpperCamelCase_=False ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase_ , map_location='cpu' )['model'] remove_ignore_keys_(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = state_dict['encoder.embed_tokens.weight'].shape[0] SCREAMING_SNAKE_CASE__ = MBartConfig.from_pretrained(UpperCamelCase_ , vocab_size=UpperCamelCase_ ) if mbart_aa and finetuned: SCREAMING_SNAKE_CASE__ = 'relu' SCREAMING_SNAKE_CASE__ = state_dict['decoder.embed_tokens.weight'] SCREAMING_SNAKE_CASE__ = MBartForConditionalGeneration(UpperCamelCase_ ) model.model.load_state_dict(UpperCamelCase_ ) if finetuned: SCREAMING_SNAKE_CASE__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """fairseq_path""", type=str, help="""bart.large, bart.large.cnn or a path to a model.pt on local filesystem.""" ) parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--hf_config""", default="""facebook/mbart-large-cc25""", type=str, help="""Which huggingface architecture to use: mbart-large""", ) parser.add_argument("""--mbart_50""", action="""store_true""", help="""whether the model is mMART-50 checkpoint""") parser.add_argument("""--finetuned""", action="""store_true""", help="""whether the model is a fine-tuned checkpoint""") __snake_case = parser.parse_args() __snake_case = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
472
1
'''simple docstring''' import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer _lowercase : List[Any] =logging.get_logger(__name__) _lowercase : Union[str, Any] ={"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _lowercase : Tuple ={ "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } _lowercase : Union[str, Any] ={ "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } _lowercase : List[str] ={ "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } _lowercase : Union[str, Any] ={ "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } _lowercase : List[Any] ={ "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } _lowercase : List[Any] ={ "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } _lowercase : Union[str, Any] ={ "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } _lowercase : Optional[Any] ={ "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } _lowercase : Dict ={ "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class _SCREAMING_SNAKE_CASE (lowercase__ ): A__ = VOCAB_FILES_NAMES A__ = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP A__ = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION A__ = DPRContextEncoderTokenizer class _SCREAMING_SNAKE_CASE (lowercase__ ): A__ = VOCAB_FILES_NAMES A__ = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP A__ = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION A__ = DPRQuestionEncoderTokenizer _lowercase : Union[str, Any] =collections.namedtuple( "DPRSpanPrediction", ["span_score", "relevance_score", "doc_id", "start_index", "end_index", "text"] ) _lowercase : int =collections.namedtuple("DPRReaderOutput", ["start_logits", "end_logits", "relevance_logits"]) _lowercase : Dict =R"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(lowercase__ ) class _SCREAMING_SNAKE_CASE : def __call__( self : List[str] , __UpperCamelCase : Any , __UpperCamelCase : Optional[str] = None , __UpperCamelCase : Optional[str] = None , __UpperCamelCase : Union[bool, str] = False , __UpperCamelCase : Union[bool, str] = False , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : Optional[Union[str, TensorType]] = None , __UpperCamelCase : Optional[bool] = None , **__UpperCamelCase : str , ) -> BatchEncoding: """simple docstring""" if titles is None and texts is None: return super().__call__( __UpperCamelCase , padding=__UpperCamelCase , truncation=__UpperCamelCase , max_length=__UpperCamelCase , return_tensors=__UpperCamelCase , return_attention_mask=__UpperCamelCase , **__UpperCamelCase , ) elif titles is None or texts is None: snake_case__ : Tuple = titles if texts is None else texts return super().__call__( __UpperCamelCase , __UpperCamelCase , padding=__UpperCamelCase , truncation=__UpperCamelCase , max_length=__UpperCamelCase , return_tensors=__UpperCamelCase , return_attention_mask=__UpperCamelCase , **__UpperCamelCase , ) snake_case__ : Union[str, Any] = titles if not isinstance(__UpperCamelCase , __UpperCamelCase ) else [titles] snake_case__ : List[str] = texts if not isinstance(__UpperCamelCase , __UpperCamelCase ) else [texts] snake_case__ : Tuple = len(__UpperCamelCase ) snake_case__ : int = questions if not isinstance(__UpperCamelCase , __UpperCamelCase ) else [questions] * n_passages assert len(__UpperCamelCase ) == len( __UpperCamelCase ), F'''There should be as many titles than texts but got {len(__UpperCamelCase )} titles and {len(__UpperCamelCase )} texts.''' snake_case__ : Union[str, Any] = super().__call__(__UpperCamelCase , __UpperCamelCase , padding=__UpperCamelCase , truncation=__UpperCamelCase )['''input_ids'''] snake_case__ : Dict = super().__call__(__UpperCamelCase , add_special_tokens=__UpperCamelCase , padding=__UpperCamelCase , truncation=__UpperCamelCase )['''input_ids'''] snake_case__ : int = { '''input_ids''': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(__UpperCamelCase , __UpperCamelCase ) ] } if return_attention_mask is not False: snake_case__ : Dict = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) snake_case__ : str = attention_mask return self.pad(__UpperCamelCase , padding=__UpperCamelCase , max_length=__UpperCamelCase , return_tensors=__UpperCamelCase ) def lowerCAmelCase ( self : Optional[Any] , __UpperCamelCase : BatchEncoding , __UpperCamelCase : DPRReaderOutput , __UpperCamelCase : int = 16 , __UpperCamelCase : int = 64 , __UpperCamelCase : int = 4 , ) -> List[DPRSpanPrediction]: """simple docstring""" snake_case__ : str = reader_input['''input_ids'''] snake_case__ , snake_case__ , snake_case__ : Union[str, Any] = reader_output[:3] snake_case__ : List[Any] = len(__UpperCamelCase ) snake_case__ : Any = sorted(range(__UpperCamelCase ) , reverse=__UpperCamelCase , key=relevance_logits.__getitem__ ) snake_case__ : List[DPRReaderOutput] = [] for doc_id in sorted_docs: snake_case__ : Any = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence snake_case__ : List[str] = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: snake_case__ : Optional[Any] = sequence_ids.index(self.pad_token_id ) else: snake_case__ : Optional[int] = len(__UpperCamelCase ) snake_case__ : Tuple = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=__UpperCamelCase , top_spans=__UpperCamelCase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=__UpperCamelCase , start_index=__UpperCamelCase , end_index=__UpperCamelCase , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(__UpperCamelCase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def lowerCAmelCase ( self : List[str] , __UpperCamelCase : List[int] , __UpperCamelCase : List[int] , __UpperCamelCase : int , __UpperCamelCase : int , ) -> List[DPRSpanPrediction]: """simple docstring""" snake_case__ : Optional[int] = [] for start_index, start_score in enumerate(__UpperCamelCase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) snake_case__ : Tuple = sorted(__UpperCamelCase , key=lambda __UpperCamelCase : x[1] , reverse=__UpperCamelCase ) snake_case__ : Tuple = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' snake_case__ : int = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(__UpperCamelCase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(lowercase__ ) class _SCREAMING_SNAKE_CASE (lowercase__, lowercase__ ): A__ = VOCAB_FILES_NAMES A__ = READER_PRETRAINED_VOCAB_FILES_MAP A__ = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ = READER_PRETRAINED_INIT_CONFIGURATION A__ = ['input_ids', 'attention_mask'] A__ = DPRReaderTokenizer
574
'''simple docstring''' import argparse import os import re _lowercase : str ="src/transformers" # Pattern that looks at the indentation in a line. _lowercase : List[Any] =re.compile(R"^(\s*)\S") # Pattern that matches `"key":" and puts `key` in group 0. _lowercase : Optional[Any] =re.compile(R"^\s*\"([^\"]+)\":") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _lowercase : Tuple =re.compile(R"^\s*_import_structure\[\"([^\"]+)\"\]") # Pattern that matches `"key",` and puts `key` in group 0. _lowercase : List[Any] =re.compile(R"^\s*\"([^\"]+)\",\s*$") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _lowercase : int =re.compile(R"\[([^\]]+)\]") def __UpperCAmelCase ( UpperCamelCase__ :List[str] ) -> Tuple: snake_case__ : str = _re_indent.search(UpperCamelCase__ ) return "" if search is None else search.groups()[0] def __UpperCAmelCase ( UpperCamelCase__ :int , UpperCamelCase__ :int="" , UpperCamelCase__ :Optional[int]=None , UpperCamelCase__ :str=None ) -> int: snake_case__ : Union[str, Any] = 0 snake_case__ : int = code.split('''\n''' ) if start_prompt is not None: while not lines[index].startswith(UpperCamelCase__ ): index += 1 snake_case__ : Dict = ['''\n'''.join(lines[:index] )] else: snake_case__ : Union[str, Any] = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). snake_case__ : int = [lines[index]] index += 1 while index < len(UpperCamelCase__ ) and (end_prompt is None or not lines[index].startswith(UpperCamelCase__ )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(UpperCamelCase__ ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ''' ''' ): current_block.append(lines[index] ) blocks.append('''\n'''.join(UpperCamelCase__ ) ) if index < len(UpperCamelCase__ ) - 1: snake_case__ : Any = [lines[index + 1]] index += 1 else: snake_case__ : Any = [] else: blocks.append('''\n'''.join(UpperCamelCase__ ) ) snake_case__ : List[str] = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(UpperCamelCase__ ) > 0: blocks.append('''\n'''.join(UpperCamelCase__ ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(UpperCamelCase__ ): blocks.append('''\n'''.join(lines[index:] ) ) return blocks def __UpperCAmelCase ( UpperCamelCase__ :Union[str, Any] ) -> int: def _inner(UpperCamelCase__ :Dict ): return key(UpperCamelCase__ ).lower().replace('''_''' , '''''' ) return _inner def __UpperCAmelCase ( UpperCamelCase__ :str , UpperCamelCase__ :List[str]=None ) -> Optional[Any]: # If no key is provided, we use a noop. def noop(UpperCamelCase__ :List[str] ): return x if key is None: snake_case__ : Optional[Any] = noop # Constants are all uppercase, they go first. snake_case__ : Dict = [obj for obj in objects if key(UpperCamelCase__ ).isupper()] # Classes are not all uppercase but start with a capital, they go second. snake_case__ : Optional[int] = [obj for obj in objects if key(UpperCamelCase__ )[0].isupper() and not key(UpperCamelCase__ ).isupper()] # Functions begin with a lowercase, they go last. snake_case__ : Any = [obj for obj in objects if not key(UpperCamelCase__ )[0].isupper()] snake_case__ : Union[str, Any] = ignore_underscore(UpperCamelCase__ ) return sorted(UpperCamelCase__ , key=UpperCamelCase__ ) + sorted(UpperCamelCase__ , key=UpperCamelCase__ ) + sorted(UpperCamelCase__ , key=UpperCamelCase__ ) def __UpperCAmelCase ( UpperCamelCase__ :List[Any] ) -> List[Any]: # This inner function sort imports between [ ]. def _replace(UpperCamelCase__ :Union[str, Any] ): snake_case__ : Union[str, Any] = match.groups()[0] if "," not in imports: return F'''[{imports}]''' snake_case__ : Dict = [part.strip().replace('''"''' , '''''' ) for part in imports.split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: snake_case__ : Tuple = keys[:-1] return "[" + ", ".join([F'''"{k}"''' for k in sort_objects(UpperCamelCase__ )] ) + "]" snake_case__ : Optional[int] = import_statement.split('''\n''' ) if len(UpperCamelCase__ ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. snake_case__ : Optional[Any] = 2 if lines[1].strip() == '''[''' else 1 snake_case__ : Union[str, Any] = [(i, _re_strip_line.search(UpperCamelCase__ ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] snake_case__ : Dict = sort_objects(UpperCamelCase__ , key=lambda UpperCamelCase__ : x[1] ) snake_case__ : Any = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(UpperCamelCase__ ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: snake_case__ : Union[str, Any] = _re_bracket_content.sub(_replace , lines[1] ) else: snake_case__ : Any = [part.strip().replace('''"''' , '''''' ) for part in lines[1].split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: snake_case__ : Dict = keys[:-1] snake_case__ : Union[str, Any] = get_indent(lines[1] ) + ''', '''.join([F'''"{k}"''' for k in sort_objects(UpperCamelCase__ )] ) return "\n".join(UpperCamelCase__ ) else: # Finally we have to deal with imports fitting on one line snake_case__ : Dict = _re_bracket_content.sub(_replace , UpperCamelCase__ ) return import_statement def __UpperCAmelCase ( UpperCamelCase__ :Optional[int] , UpperCamelCase__ :Dict=True ) -> Dict: with open(UpperCamelCase__ , encoding='''utf-8''' ) as f: snake_case__ : Optional[int] = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 snake_case__ : Optional[Any] = split_code_in_indented_blocks( UpperCamelCase__ , start_prompt='''_import_structure = {''' , end_prompt='''if TYPE_CHECKING:''' ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(UpperCamelCase__ ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. snake_case__ : Any = main_blocks[block_idx] snake_case__ : Dict = block.split('''\n''' ) # Get to the start of the imports. snake_case__ : List[str] = 0 while line_idx < len(UpperCamelCase__ ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: snake_case__ : List[Any] = len(UpperCamelCase__ ) else: line_idx += 1 if line_idx >= len(UpperCamelCase__ ): continue # Ignore beginning and last line: they don't contain anything. snake_case__ : Optional[Any] = '''\n'''.join(block_lines[line_idx:-1] ) snake_case__ : Dict = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. snake_case__ : List[str] = split_code_in_indented_blocks(UpperCamelCase__ , indent_level=UpperCamelCase__ ) # We have two categories of import key: list or _import_structure[key].append/extend snake_case__ : Tuple = _re_direct_key if '''_import_structure = {''' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. snake_case__ : Optional[int] = [(pattern.search(UpperCamelCase__ ).groups()[0] if pattern.search(UpperCamelCase__ ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. snake_case__ : Any = [(i, key) for i, key in enumerate(UpperCamelCase__ ) if key is not None] snake_case__ : Optional[int] = [x[0] for x in sorted(UpperCamelCase__ , key=lambda UpperCamelCase__ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. snake_case__ : Dict = 0 snake_case__ : Dict = [] for i in range(len(UpperCamelCase__ ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: snake_case__ : Tuple = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(UpperCamelCase__ ) count += 1 # And we put our main block back together with its first and last line. snake_case__ : int = '''\n'''.join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(UpperCamelCase__ ): if check_only: return True else: print(F'''Overwriting {file}.''' ) with open(UpperCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write('''\n'''.join(UpperCamelCase__ ) ) def __UpperCAmelCase ( UpperCamelCase__ :Optional[Any]=True ) -> Union[str, Any]: snake_case__ : str = [] for root, _, files in os.walk(UpperCamelCase__ ): if "__init__.py" in files: snake_case__ : int = sort_imports(os.path.join(UpperCamelCase__ , '''__init__.py''' ) , check_only=UpperCamelCase__ ) if result: snake_case__ : Optional[Any] = [os.path.join(UpperCamelCase__ , '''__init__.py''' )] if len(UpperCamelCase__ ) > 0: raise ValueError(F'''Would overwrite {len(UpperCamelCase__ )} files, run `make style`.''' ) if __name__ == "__main__": _lowercase : Optional[Any] =argparse.ArgumentParser() parser.add_argument("--check_only", action="store_true", help="Whether to only check or fix style.") _lowercase : Tuple =parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
574
1
"""simple docstring""" import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase = '▁' UpperCamelCase = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class UpperCamelCase__ ( snake_case_ , unittest.TestCase ): """simple docstring""" A__ : List[Any] = BertGenerationTokenizer A__ : str = False A__ : str = True def snake_case__ ( self ) -> Tuple: super().setUp() A__ = BertGenerationTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ) -> List[str]: A__ = "<s>" A__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ) -> List[str]: A__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "<pad>" ) self.assertEqual(len(lowerCamelCase__ ) , 1002 ) def snake_case__ ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def snake_case__ ( self ) -> Optional[Any]: A__ = BertGenerationTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) A__ = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCamelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [285, 46, 10, 170, 382] , ) A__ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) A__ = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) A__ = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def snake_case__ ( self ) -> Tuple: return BertGenerationTokenizer.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) @slow def snake_case__ ( self ) -> str: A__ = "Hello World!" A__ = [18536, 2260, 101] self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ) -> Any: A__ = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) A__ = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @require_torch @slow def snake_case__ ( self ) -> Tuple: import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence A__ = list(self.big_tokenizer.get_vocab().keys() )[:10] A__ = " ".join(lowerCamelCase__ ) A__ = self.big_tokenizer.encode_plus(lowerCamelCase__ , return_tensors="pt" , return_token_type_ids=lowerCamelCase__ ) A__ = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=lowerCamelCase__ ) A__ = BertGenerationConfig() A__ = BertGenerationEncoder(lowerCamelCase__ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowerCamelCase__ ) model(**lowerCamelCase__ ) @slow def snake_case__ ( self ) -> Tuple: A__ = {"input_ids": [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name="google/bert_for_seq_generation_L-24_bbc_encoder" , revision="c817d1fd1be2ffa69431227a1fe320544943d4db" , )
104
'''simple docstring''' def __snake_case ( _UpperCAmelCase : int): UpperCamelCase = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
212
0
'''simple docstring''' import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class _lowerCamelCase : '''simple docstring''' def __init__( self , __lowercase , __lowercase = 13 , __lowercase = 64 , __lowercase = 2 , __lowercase = 3 , __lowercase = 3 , __lowercase = True , __lowercase = True , __lowercase = 128 , __lowercase=[16, 32, 64, 128] , __lowercase = 7 , __lowercase = 4 , __lowercase = 37 , __lowercase = "gelu" , __lowercase = 0.1 , __lowercase = 0.1 , __lowercase = 10 , __lowercase = 0.0_2 , __lowercase = 2 , __lowercase = 1 , __lowercase = 128 , __lowercase = [2, 2, 2, 2] , __lowercase = 2 , __lowercase = 2 , ): """simple docstring""" __A : Optional[Any] = parent __A : str = batch_size __A : Optional[Any] = image_size __A : str = patch_size __A : Tuple = num_channels __A : Union[str, Any] = is_training __A : Any = use_labels __A : Union[str, Any] = hidden_size __A : Union[str, Any] = num_hidden_layers __A : Dict = num_attention_heads __A : Dict = intermediate_size __A : List[str] = hidden_act __A : Union[str, Any] = hidden_dropout_prob __A : Union[str, Any] = attention_probs_dropout_prob __A : Optional[Any] = type_sequence_label_size __A : int = initializer_range __A : Any = encoder_stride __A : Tuple = num_attention_outputs __A : Any = embed_dim __A : Any = embed_dim + 1 __A : Dict = resolution __A : Union[str, Any] = depths __A : int = hidden_sizes __A : Optional[int] = dim __A : Tuple = mlp_expansion_ratio def snake_case__ ( self ): """simple docstring""" __A : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __A : List[str] = None if self.use_labels: __A : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __A : List[str] = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): """simple docstring""" return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def snake_case__ ( self , __lowercase , __lowercase , __lowercase ): """simple docstring""" __A : Dict = TFEfficientFormerModel(config=__lowercase ) __A : List[Any] = model(__lowercase , training=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , __lowercase , __lowercase , __lowercase ): """simple docstring""" __A : Any = self.type_sequence_label_size __A : str = TFEfficientFormerForImageClassification(__lowercase ) __A : Any = model(__lowercase , labels=__lowercase , training=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __A : List[str] = 1 __A : Any = TFEfficientFormerForImageClassification(__lowercase ) __A : Optional[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __A : Optional[Any] = model(__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self ): """simple docstring""" __A : int = self.prepare_config_and_inputs() __A : Dict = config_and_inputs __A : Optional[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class _lowerCamelCase ( snake_case_ , snake_case_ , unittest.TestCase ): '''simple docstring''' __lowercase : int = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) __lowercase : Optional[int] = ( { '''feature-extraction''': TFEfficientFormerModel, '''image-classification''': ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) __lowercase : Any = False __lowercase : int = False __lowercase : Dict = False __lowercase : Dict = False __lowercase : Optional[Any] = False def snake_case__ ( self ): """simple docstring""" __A : str = TFEfficientFormerModelTester(self ) __A : Any = ConfigTester( self , config_class=__lowercase , has_text_modality=__lowercase , hidden_size=37 ) def snake_case__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='EfficientFormer does not use inputs_embeds' ) def snake_case__ ( self ): """simple docstring""" pass @unittest.skip(reason='EfficientFormer does not support input and output embeddings' ) def snake_case__ ( self ): """simple docstring""" pass def snake_case__ ( self ): """simple docstring""" __A : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __A : Dict = model_class(__lowercase ) __A : Any = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __A : List[str] = [*signature.parameters.keys()] __A : int = ['pixel_values'] self.assertListEqual(arg_names[:1] , __lowercase ) def snake_case__ ( self ): """simple docstring""" def check_hidden_states_output(__lowercase , __lowercase , __lowercase ): __A : List[Any] = model_class(__lowercase ) __A : Tuple = model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) __A : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __A : Any = getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__lowercase ) , __lowercase ) if hasattr(self.model_tester , 'encoder_seq_length' ): __A : str = self.model_tester.encoder_seq_length if hasattr(self.model_tester , 'chunk_length' ) and self.model_tester.chunk_length > 1: __A : Optional[int] = seq_length * self.model_tester.chunk_length else: __A : Tuple = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __A : List[Any] = outputs.decoder_hidden_states self.asseretIsInstance(__lowercase , (list, tuple) ) self.assertEqual(len(__lowercase ) , __lowercase ) __A : Dict = getattr(self.model_tester , 'seq_length' , __lowercase ) __A : Dict = getattr(self.model_tester , 'decoder_seq_length' , __lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __A : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __A : int = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __A : int = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) def snake_case__ ( self , __lowercase , __lowercase , __lowercase=False ): """simple docstring""" __A : List[str] = super()._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def snake_case__ ( self ): """simple docstring""" __A : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) @unittest.skip(reason='EfficientFormer does not implement masked image modeling yet' ) def snake_case__ ( self ): """simple docstring""" __A : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowercase ) def snake_case__ ( self ): """simple docstring""" __A : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) @slow def snake_case__ ( self ): """simple docstring""" for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __A : str = TFEfficientFormerModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def snake_case__ ( self ): """simple docstring""" __A : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() __A : List[Any] = True __A : Dict = getattr(self.model_tester , 'seq_length' , __lowercase ) __A : Any = getattr(self.model_tester , 'encoder_seq_length' , __lowercase ) __A : Union[str, Any] = getattr(self.model_tester , 'key_length' , __lowercase ) __A : Optional[int] = getattr(self.model_tester , 'chunk_length' , __lowercase ) if chunk_length is not None and hasattr(self.model_tester , 'num_hashes' ): __A : int = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __A : Any = True __A : Dict = False __A : int = True __A : Optional[Any] = model_class(__lowercase ) __A : Dict = model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) __A : Optional[Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __A : List[str] = True __A : Union[str, Any] = model_class(__lowercase ) __A : List[str] = model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) __A : Union[str, Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def snake_case__ ( self ): """simple docstring""" __A : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __A : Union[str, Any] = model_class(__lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __A : Any = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=__lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __A : Dict = model(__lowercase ) self.assertTrue(outputs_dict is not None ) def _lowercase ( ): __A : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): """simple docstring""" return ( EfficientFormerImageProcessor.from_pretrained('snap-research/efficientformer-l1-300' ) if is_vision_available() else None ) @slow def snake_case__ ( self ): """simple docstring""" __A : List[str] = TFEfficientFormerForImageClassification.from_pretrained('snap-research/efficientformer-l1-300' ) __A : List[str] = self.default_image_processor __A : List[Any] = prepare_img() __A : Union[str, Any] = image_processor(images=__lowercase , return_tensors='tf' ) # forward pass __A : List[str] = model(**__lowercase , training=__lowercase ) # verify the logits __A : Dict = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowercase ) __A : Optional[int] = tf.constant([-0.0_5_5_5, 0.4_8_2_5, -0.0_8_5_2] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowercase , atol=1E-4 ) ) @slow def snake_case__ ( self ): """simple docstring""" __A : Optional[Any] = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( 'snap-research/efficientformer-l1-300' ) __A : Tuple = self.default_image_processor __A : int = prepare_img() __A : List[str] = image_processor(images=__lowercase , return_tensors='tf' ) # forward pass __A : int = model(**__lowercase , training=__lowercase ) # verify the logits __A : Any = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowercase ) __A : int = tf.constant([-0.1_3_1_2, 0.4_3_5_3, -1.0_4_9_9] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowercase , atol=1E-4 ) )
717
'''simple docstring''' from math import pi, sqrt, tan def _lowercase ( UpperCamelCase__ : float ): if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float, UpperCamelCase__ : float ): if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def _lowercase ( UpperCamelCase__ : float ): if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def _lowercase ( UpperCamelCase__ : float ): if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float, UpperCamelCase__ : float ): if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) __A : Optional[Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(UpperCamelCase__, 2 ) * torus_radius * tube_radius def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def _lowercase ( UpperCamelCase__ : float ): if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float, UpperCamelCase__ : float ): if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) __A : int = (sidea + sidea + sidea) / 2 __A : List[Any] = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float, UpperCamelCase__ : float ): if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def _lowercase ( UpperCamelCase__ : float ): if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def _lowercase ( UpperCamelCase__ : float, UpperCamelCase__ : float ): if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def _lowercase ( UpperCamelCase__ : int, UpperCamelCase__ : float ): if not isinstance(UpperCamelCase__, UpperCamelCase__ ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(f'''Rectangle: {area_rectangle(1_0, 2_0) = }''') print(f'''Square: {area_square(1_0) = }''') print(f'''Triangle: {area_triangle(1_0, 1_0) = }''') print(f'''Triangle: {area_triangle_three_sides(5, 1_2, 1_3) = }''') print(f'''Parallelogram: {area_parallelogram(1_0, 2_0) = }''') print(f'''Rhombus: {area_rhombus(1_0, 2_0) = }''') print(f'''Trapezium: {area_trapezium(1_0, 2_0, 3_0) = }''') print(f'''Circle: {area_circle(2_0) = }''') print(f'''Ellipse: {area_ellipse(1_0, 2_0) = }''') print('\nSurface Areas of various geometric shapes: \n') print(f'''Cube: {surface_area_cube(2_0) = }''') print(f'''Cuboid: {surface_area_cuboid(1_0, 2_0, 3_0) = }''') print(f'''Sphere: {surface_area_sphere(2_0) = }''') print(f'''Hemisphere: {surface_area_hemisphere(2_0) = }''') print(f'''Cone: {surface_area_cone(1_0, 2_0) = }''') print(f'''Conical Frustum: {surface_area_conical_frustum(1_0, 2_0, 3_0) = }''') print(f'''Cylinder: {surface_area_cylinder(1_0, 2_0) = }''') print(f'''Torus: {surface_area_torus(2_0, 1_0) = }''') print(f'''Equilateral Triangle: {area_reg_polygon(3, 1_0) = }''') print(f'''Square: {area_reg_polygon(4, 1_0) = }''') print(f'''Reqular Pentagon: {area_reg_polygon(5, 1_0) = }''')
540
0