code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging lowercase_ = logging.get_logger(__name__) class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''linear''' lowerCAmelCase_ = '''cosine''' lowerCAmelCase_ = '''cosine_with_restarts''' lowerCAmelCase_ = '''polynomial''' lowerCAmelCase_ = '''constant''' lowerCAmelCase_ = '''constant_with_warmup''' lowerCAmelCase_ = '''piecewise_constant''' def a__ ( snake_case , snake_case = -1 ): """simple docstring""" return LambdaLR(snake_case , lambda snake_case : 1 , last_epoch=snake_case ) def a__ ( snake_case , snake_case , snake_case = -1 ): """simple docstring""" def lr_lambda(snake_case ): if current_step < num_warmup_steps: return float(snake_case ) / float(max(1.0 , snake_case ) ) return 1.0 return LambdaLR(snake_case , snake_case , last_epoch=snake_case ) def a__ ( snake_case , snake_case , snake_case = -1 ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = {} __SCREAMING_SNAKE_CASE : Tuple = step_rules.split(''',''' ) for rule_str in rule_list[:-1]: __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : List[Any] = rule_str.split(''':''' ) __SCREAMING_SNAKE_CASE : List[Any] = int(snake_case ) __SCREAMING_SNAKE_CASE : Dict = float(snake_case ) __SCREAMING_SNAKE_CASE : str = value __SCREAMING_SNAKE_CASE : Optional[int] = float(rule_list[-1] ) def create_rules_function(snake_case , snake_case ): def rule_func(snake_case ) -> float: __SCREAMING_SNAKE_CASE : Optional[int] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(snake_case ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __SCREAMING_SNAKE_CASE : Union[str, Any] = create_rules_function(snake_case , snake_case ) return LambdaLR(snake_case , snake_case , last_epoch=snake_case ) def a__ ( snake_case , snake_case , snake_case , snake_case=-1 ): """simple docstring""" def lr_lambda(snake_case ): if current_step < num_warmup_steps: return float(snake_case ) / float(max(1 , snake_case ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(snake_case , snake_case , snake_case ) def a__ ( snake_case , snake_case , snake_case , snake_case = 0.5 , snake_case = -1 ): """simple docstring""" def lr_lambda(snake_case ): if current_step < num_warmup_steps: return float(snake_case ) / float(max(1 , snake_case ) ) __SCREAMING_SNAKE_CASE : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(snake_case ) * 2.0 * progress )) ) return LambdaLR(snake_case , snake_case , snake_case ) def a__ ( snake_case , snake_case , snake_case , snake_case = 1 , snake_case = -1 ): """simple docstring""" def lr_lambda(snake_case ): if current_step < num_warmup_steps: return float(snake_case ) / float(max(1 , snake_case ) ) __SCREAMING_SNAKE_CASE : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(snake_case ) * progress) % 1.0) )) ) return LambdaLR(snake_case , snake_case , snake_case ) def a__ ( snake_case , snake_case , snake_case , snake_case=1E-7 , snake_case=1.0 , snake_case=-1 ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = optimizer.defaults['''lr'''] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''' ) def lr_lambda(snake_case ): if current_step < num_warmup_steps: return float(snake_case ) / float(max(1 , snake_case ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __SCREAMING_SNAKE_CASE : Any = lr_init - lr_end __SCREAMING_SNAKE_CASE : List[str] = num_training_steps - num_warmup_steps __SCREAMING_SNAKE_CASE : Union[str, Any] = 1 - (current_step - num_warmup_steps) / decay_steps __SCREAMING_SNAKE_CASE : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(snake_case , snake_case , snake_case ) lowercase_ = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def a__ ( snake_case , snake_case , snake_case = None , snake_case = None , snake_case = None , snake_case = 1 , snake_case = 1.0 , snake_case = -1 , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = SchedulerType(snake_case ) __SCREAMING_SNAKE_CASE : Any = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(snake_case , last_epoch=snake_case ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(snake_case , step_rules=snake_case , last_epoch=snake_case ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''' ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(snake_case , num_warmup_steps=snake_case , last_epoch=snake_case ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''' ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( snake_case , num_warmup_steps=snake_case , num_training_steps=snake_case , num_cycles=snake_case , last_epoch=snake_case , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( snake_case , num_warmup_steps=snake_case , num_training_steps=snake_case , power=snake_case , last_epoch=snake_case , ) return schedule_func( snake_case , num_warmup_steps=snake_case , num_training_steps=snake_case , last_epoch=snake_case )
303
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class UpperCamelCase : def __init__( self, lowerCAmelCase__) -> Optional[int]: snake_case_ = data snake_case_ = None class UpperCamelCase : def __init__( self) -> Dict: snake_case_ = None snake_case_ = None def __iter__( self) -> Iterator[Any]: snake_case_ = self.head while self.head: yield node.data snake_case_ = node.next if node == self.head: break def __len__( self) -> int: return sum(1 for _ in self) def __repr__( self) -> str: return "->".join(str(lowerCAmelCase__) for item in iter(self)) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(len(self), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(0, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> None: if index < 0 or index > len(self): raise IndexError('list index out of range.') snake_case_ = Node(lowerCAmelCase__) if self.head is None: snake_case_ = new_node # first node points itself snake_case_ = snake_case_ = new_node elif index == 0: # insert at head snake_case_ = self.head snake_case_ = snake_case_ = new_node else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = new_node if index == len(self) - 1: # insert at tail snake_case_ = new_node def a_ ( self) -> str: return self.delete_nth(0) def a_ ( self) -> Any: return self.delete_nth(len(self) - 1) def a_ ( self, lowerCAmelCase__ = 0) -> Any: if not 0 <= index < len(self): raise IndexError('list index out of range.') snake_case_ = self.head if self.head == self.tail: # just one node snake_case_ = snake_case_ = None elif index == 0: # delete head node snake_case_ = self.tail.next.next snake_case_ = self.head.next else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = temp.next.next if index == len(self) - 1: # delete at tail snake_case_ = temp return delete_node.data def a_ ( self) -> bool: return len(self) == 0 def UpperCAmelCase ( ) -> None: snake_case_ = CircularLinkedList() assert len(UpperCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(UpperCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(UpperCAmelCase ) == i circular_linked_list.insert_nth(UpperCAmelCase , i + 1 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
69
0
"""simple docstring""" def __magic_name__ ( __snake_case : Union[str, Any] = 50 ) -> int: lowercase : str = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"{solution() = }")
202
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCamelCase = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __UpperCamelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) __UpperCamelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __UpperCamelCase = re.compile(r'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __UpperCamelCase = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def UpperCAmelCase ( UpperCAmelCase ) -> List[Any]: snake_case_ = None # source code of `config_class` snake_case_ = inspect.getsource(UpperCAmelCase ) snake_case_ = _re_checkpoint.findall(UpperCAmelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): snake_case_ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: snake_case_ = ckpt_name break return checkpoint def UpperCAmelCase ( ) -> Union[str, Any]: snake_case_ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ = get_checkpoint_from_config_class(UpperCAmelCase ) snake_case_ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = '\n'.join(sorted(UpperCAmelCase ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer lowercase : Optional[Any] = logging.get_logger(__name__) lowercase : Dict = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase : Optional[Any] = { """vocab_file""": { """allegro/herbert-base-cased""": """https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json""" }, """merges_file""": { """allegro/herbert-base-cased""": """https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt""" }, } lowercase : List[str] = {"""allegro/herbert-base-cased""": 5_1_4} lowercase : str = {} class A__ ( lowerCAmelCase__ ): """simple docstring""" __A : List[Any] = VOCAB_FILES_NAMES __A : List[str] = PRETRAINED_VOCAB_FILES_MAP __A : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION __A : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : List[str] = HerbertTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase="</s>" , **lowercase , ) -> int: '''simple docstring''' super().__init__( lowerCAmelCase__ , lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) def __lowercase ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' a__ : Tuple = [self.cls_token_id] a__ : List[str] = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __lowercase ( self , lowercase , lowercase = None , lowercase = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__)) + [1] return [1] + ([0] * len(lowerCAmelCase__)) + [1] + ([0] * len(lowerCAmelCase__)) + [1] def __lowercase ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' a__ : str = [self.sep_token_id] a__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def __lowercase ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' a__ : List[Any] = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__) return tuple(lowerCAmelCase__)
99
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __UpperCamelCase = 0 __UpperCamelCase = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __UpperCamelCase = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __UpperCamelCase = tuple[int, int] class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, ) -> None: snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = g_cost snake_case_ = parent snake_case_ = self.calculate_heuristic() snake_case_ = self.g_cost + self.h_cost def a_ ( self) -> float: snake_case_ = self.pos_x - self.goal_x snake_case_ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase__) + abs(lowerCAmelCase__) else: return sqrt(dy**2 + dx**2) def __lt__( self, lowerCAmelCase__) -> bool: return self.f_cost < other.f_cost class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = Node(start[1], start[0], goal[1], goal[0], 0, lowerCAmelCase__) snake_case_ = Node(goal[1], goal[0], goal[1], goal[0], 9_9999, lowerCAmelCase__) snake_case_ = [self.start] snake_case_ = [] snake_case_ = False def a_ ( self) -> list[TPosition]: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() snake_case_ = self.open_nodes.pop(0) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase__) self.closed_nodes.append(lowerCAmelCase__) snake_case_ = self.get_successors(lowerCAmelCase__) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__) else: self.open_nodes.append(lowerCAmelCase__) return [self.start.pos] def a_ ( self, lowerCAmelCase__) -> list[Node]: snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(lowerCAmelCase__) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__, lowerCAmelCase__, self.target.pos_y, self.target.pos_x, parent.g_cost + 1, lowerCAmelCase__, )) return successors def a_ ( self, lowerCAmelCase__) -> list[TPosition]: snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x)) snake_case_ = current_node.parent path.reverse() return path class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> None: snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = False def a_ ( self) -> list[TPosition]: while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() snake_case_ = self.fwd_astar.open_nodes.pop(0) snake_case_ = self.bwd_astar.open_nodes.pop(0) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase__, lowerCAmelCase__) self.fwd_astar.closed_nodes.append(lowerCAmelCase__) self.bwd_astar.closed_nodes.append(lowerCAmelCase__) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase__), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase__), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase__) else: astar.open_nodes.append(lowerCAmelCase__) return [self.fwd_astar.start.pos] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> list[TPosition]: snake_case_ = self.fwd_astar.retrace_path(lowerCAmelCase__) snake_case_ = self.bwd_astar.retrace_path(lowerCAmelCase__) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __UpperCamelCase = (0, 0) __UpperCamelCase = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __UpperCamelCase = time.time() __UpperCamelCase = AStar(init, goal) __UpperCamelCase = a_star.search() __UpperCamelCase = time.time() - start_time print(F"""AStar execution time = {end_time:f} seconds""") __UpperCamelCase = time.time() __UpperCamelCase = BidirectionalAStar(init, goal) __UpperCamelCase = time.time() - bd_start_time print(F"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
69
0
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin A_ : Optional[Any] = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class A_ ( lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' a__ = BartphoTokenizer a__ = False a__ = True def lowerCAmelCase_ (self ) -> int: super().setUp() __UpperCAmelCase = ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] __UpperCAmelCase = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) __UpperCAmelCase = {'''unk_token''': '''<unk>'''} __UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''monolingual_vocab_file'''] ) with open(self.monolingual_vocab_file , '''w''' , encoding='''utf-8''' ) as fp: for token in vocab_tokens: fp.write(F'''{token} {vocab_tokens[token]}\n''' ) __UpperCAmelCase = BartphoTokenizer(lowerCAmelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase_ (self , **lowercase__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def lowerCAmelCase_ (self , lowercase__ ) -> Union[str, Any]: __UpperCAmelCase = '''This is a là test''' __UpperCAmelCase = '''This is a<unk><unk> test''' return input_text, output_text def lowerCAmelCase_ (self ) -> Tuple: __UpperCAmelCase = BartphoTokenizer(lowerCAmelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) __UpperCAmelCase = '''This is a là test''' __UpperCAmelCase = '''▁This ▁is ▁a ▁l à ▁t est'''.split() __UpperCAmelCase = tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) __UpperCAmelCase = tokens + [tokenizer.unk_token] __UpperCAmelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , lowerCAmelCase__ )
333
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: while a != 0: snake_case_ , snake_case_ = b % a, a return b def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: if gcd(UpperCAmelCase , UpperCAmelCase ) != 1: snake_case_ = f'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = 1, 0, a snake_case_ , snake_case_ , snake_case_ = 0, 1, m while va != 0: snake_case_ = ua // va snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
69
0
from functools import reduce _snake_case : Tuple = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def lowerCAmelCase_ ( __lowerCamelCase = N ): return max( # mypy cannot properly interpret reduce int(reduce(lambda __lowerCamelCase , __lowerCamelCase : str(int(__lowerCamelCase ) * int(__lowerCamelCase ) ) , n[i : i + 1_3] ) ) for i in range(len(__lowerCamelCase ) - 1_2 ) ) if __name__ == "__main__": print(f'''{solution() = }''')
123
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase = { '''configuration_bert''': ['''BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BertConfig''', '''BertOnnxConfig'''], '''tokenization_bert''': ['''BasicTokenizer''', '''BertTokenizer''', '''WordpieceTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''BertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BertForMaskedLM''', '''BertForMultipleChoice''', '''BertForNextSentencePrediction''', '''BertForPreTraining''', '''BertForQuestionAnswering''', '''BertForSequenceClassification''', '''BertForTokenClassification''', '''BertLayer''', '''BertLMHeadModel''', '''BertModel''', '''BertPreTrainedModel''', '''load_tf_weights_in_bert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBertEmbeddings''', '''TFBertForMaskedLM''', '''TFBertForMultipleChoice''', '''TFBertForNextSentencePrediction''', '''TFBertForPreTraining''', '''TFBertForQuestionAnswering''', '''TFBertForSequenceClassification''', '''TFBertForTokenClassification''', '''TFBertLMHeadModel''', '''TFBertMainLayer''', '''TFBertModel''', '''TFBertPreTrainedModel''', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFBertTokenizer'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxBertForCausalLM''', '''FlaxBertForMaskedLM''', '''FlaxBertForMultipleChoice''', '''FlaxBertForNextSentencePrediction''', '''FlaxBertForPreTraining''', '''FlaxBertForQuestionAnswering''', '''FlaxBertForSequenceClassification''', '''FlaxBertForTokenClassification''', '''FlaxBertModel''', '''FlaxBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean SCREAMING_SNAKE_CASE__ : Optional[Any] = 0 SCREAMING_SNAKE_CASE__ : int = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] SCREAMING_SNAKE_CASE__ : Tuple = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right SCREAMING_SNAKE_CASE__ : Optional[int] = tuple[int, int] class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ) -> None: lowerCamelCase : str = pos_x lowerCamelCase : List[Any] = pos_y lowerCamelCase : Optional[int] = (pos_y, pos_x) lowerCamelCase : int = goal_x lowerCamelCase : Tuple = goal_y lowerCamelCase : int = g_cost lowerCamelCase : List[Any] = parent lowerCamelCase : Union[str, Any] = self.calculate_heuristic() lowerCamelCase : List[Any] = self.g_cost + self.h_cost def _lowercase ( self ) -> float: lowerCamelCase : Tuple = self.pos_x - self.goal_x lowerCamelCase : Any = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase__ ) + abs(lowerCAmelCase__ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , UpperCamelCase__ ) -> bool: return self.f_cost < other.f_cost class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ , UpperCamelCase__ ) -> Union[str, Any]: lowerCamelCase : Any = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCAmelCase__ ) lowerCamelCase : str = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowerCAmelCase__ ) lowerCamelCase : List[Any] = [self.start] lowerCamelCase : List[str] = [] lowerCamelCase : Tuple = False def _lowercase ( self ) -> list[TPosition]: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() lowerCamelCase : Optional[int] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase__ ) self.closed_nodes.append(lowerCAmelCase__ ) lowerCamelCase : Any = self.get_successors(lowerCAmelCase__ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__ ) else: # retrieve the best current path lowerCamelCase : str = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__ ) else: self.open_nodes.append(lowerCAmelCase__ ) return [self.start.pos] def _lowercase ( self , UpperCamelCase__ ) -> list[Node]: lowerCamelCase : Optional[Any] = [] for action in delta: lowerCamelCase : str = parent.pos_x + action[1] lowerCamelCase : Dict = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCAmelCase__ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__ , lowerCAmelCase__ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCAmelCase__ , ) ) return successors def _lowercase ( self , UpperCamelCase__ ) -> list[TPosition]: lowerCamelCase : Any = node lowerCamelCase : str = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) lowerCamelCase : Optional[int] = current_node.parent path.reverse() return path class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ , UpperCamelCase__ ) -> None: lowerCamelCase : Tuple = AStar(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCamelCase : Union[str, Any] = AStar(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCamelCase : Dict = False def _lowercase ( self ) -> list[TPosition]: while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() lowerCamelCase : Optional[int] = self.fwd_astar.open_nodes.pop(0 ) lowerCamelCase : str = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase__ , lowerCAmelCase__ ) self.fwd_astar.closed_nodes.append(lowerCAmelCase__ ) self.bwd_astar.closed_nodes.append(lowerCAmelCase__ ) lowerCamelCase : Tuple = current_bwd_node lowerCamelCase : str = current_fwd_node lowerCamelCase : List[str] = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase__ ), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase__ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase__ ) else: # retrieve the best current path lowerCamelCase : Dict = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase__ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase__ ) else: astar.open_nodes.append(lowerCAmelCase__ ) return [self.fwd_astar.start.pos] def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> list[TPosition]: lowerCamelCase : Optional[int] = self.fwd_astar.retrace_path(lowerCAmelCase__ ) lowerCamelCase : List[Any] = self.bwd_astar.retrace_path(lowerCAmelCase__ ) bwd_path.pop() bwd_path.reverse() lowerCamelCase : str = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] SCREAMING_SNAKE_CASE__ : List[Any] = (0, 0) SCREAMING_SNAKE_CASE__ : Tuple = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) SCREAMING_SNAKE_CASE__ : Union[str, Any] = time.time() SCREAMING_SNAKE_CASE__ : Optional[Any] = AStar(init, goal) SCREAMING_SNAKE_CASE__ : List[str] = a_star.search() SCREAMING_SNAKE_CASE__ : Optional[Any] = time.time() - start_time print(f'''AStar execution time = {end_time:f} seconds''') SCREAMING_SNAKE_CASE__ : Dict = time.time() SCREAMING_SNAKE_CASE__ : List[Any] = BidirectionalAStar(init, goal) SCREAMING_SNAKE_CASE__ : Any = time.time() - bd_start_time print(f'''BidirectionalAStar execution time = {bd_end_time:f} seconds''')
48
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.linear_k''': '''encoder.layers.*.self_attn.linear_k''', '''self_attn.linear_v''': '''encoder.layers.*.self_attn.linear_v''', '''self_attn.linear_q''': '''encoder.layers.*.self_attn.linear_q''', '''self_attn.pos_bias_u''': '''encoder.layers.*.self_attn.pos_bias_u''', '''self_attn.pos_bias_v''': '''encoder.layers.*.self_attn.pos_bias_v''', '''self_attn.linear_out''': '''encoder.layers.*.self_attn.linear_out''', '''self_attn.linear_pos''': '''encoder.layers.*.self_attn.linear_pos''', '''self_attn.rotary_emb''': '''encoder.embed_positions''', '''self_attn_layer_norm''': '''encoder.layers.*.self_attn_layer_norm''', '''conv_module.pointwise_conv1''': '''encoder.layers.*.conv_module.pointwise_conv1''', '''conv_module.pointwise_conv2''': '''encoder.layers.*.conv_module.pointwise_conv2''', '''conv_module.depthwise_conv''': '''encoder.layers.*.conv_module.depthwise_conv''', '''conv_module.batch_norm''': '''encoder.layers.*.conv_module.batch_norm''', '''conv_module.layer_norm''': '''encoder.layers.*.conv_module.layer_norm''', '''ffn1.w_1''': '''encoder.layers.*.ffn1.intermediate_dense''', '''ffn1.w_2''': '''encoder.layers.*.ffn1.output_dense''', '''ffn1.layer_norm''': '''encoder.layers.*.ffn1_layer_norm''', '''ffn2.w_1''': '''encoder.layers.*.ffn2.intermediate_dense''', '''ffn2.w_2''': '''encoder.layers.*.ffn2.output_dense''', '''ffn2.layer_norm''': '''encoder.layers.*.ffn2_layer_norm''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __UpperCamelCase = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: for attribute in key.split('.' ): snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ) if weight_type is not None: snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ).shape else: snake_case_ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": snake_case_ = value elif weight_type == "weight_g": snake_case_ = value elif weight_type == "weight_v": snake_case_ = value elif weight_type == "bias": snake_case_ = value elif weight_type == "running_mean": snake_case_ = value elif weight_type == "running_var": snake_case_ = value elif weight_type == "num_batches_tracked": snake_case_ = value elif weight_type == "inv_freq": snake_case_ = value else: snake_case_ = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = [] snake_case_ = fairseq_model.state_dict() snake_case_ = hf_model.wavaveca_conformer.feature_extractor for name, value in fairseq_dict.items(): snake_case_ = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) snake_case_ = True else: for key, mapped_key in MAPPING.items(): snake_case_ = 'wav2vec2_conformer.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case_ = True if "*" in mapped_key: snake_case_ = name.split(UpperCAmelCase )[0].split('.' )[-2] snake_case_ = mapped_key.replace('*' , UpperCAmelCase ) if "pos_bias_u" in name: snake_case_ = None elif "pos_bias_v" in name: snake_case_ = None elif "weight_g" in name: snake_case_ = 'weight_g' elif "weight_v" in name: snake_case_ = 'weight_v' elif "bias" in name: snake_case_ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case_ = 'weight' elif "running_mean" in name: snake_case_ = 'running_mean' elif "inv_freq" in name: snake_case_ = 'inv_freq' elif "running_var" in name: snake_case_ = 'running_var' elif "num_batches_tracked" in name: snake_case_ = 'num_batches_tracked' else: snake_case_ = None set_recursively(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) continue if not is_used: unused_weights.append(UpperCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: snake_case_ = full_name.split('conv_layers.' )[-1] snake_case_ = name.split('.' ) snake_case_ = int(items[0] ) snake_case_ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase ) @torch.no_grad() def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=True ) -> str: if config_path is not None: snake_case_ = WavaVecaConformerConfig.from_pretrained(UpperCAmelCase , hidden_act='swish' ) else: snake_case_ = WavaVecaConformerConfig() if "rope" in checkpoint_path: snake_case_ = 'rotary' if is_finetuned: if dict_path: snake_case_ = Dictionary.load(UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq snake_case_ = target_dict.pad_index snake_case_ = target_dict.bos_index snake_case_ = target_dict.eos_index snake_case_ = len(target_dict.symbols ) snake_case_ = os.path.join(UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase ) ) return os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) snake_case_ = target_dict.indices # fairseq has the <pad> and <s> switched snake_case_ = 0 snake_case_ = 1 with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(UpperCAmelCase , UpperCAmelCase ) snake_case_ = WavaVecaCTCTokenizer( UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase , ) snake_case_ = True if config.feat_extract_norm == 'layer' else False snake_case_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=UpperCAmelCase , return_attention_mask=UpperCAmelCase , ) snake_case_ = WavaVecaProcessor(feature_extractor=UpperCAmelCase , tokenizer=UpperCAmelCase ) processor.save_pretrained(UpperCAmelCase ) snake_case_ = WavaVecaConformerForCTC(UpperCAmelCase ) else: snake_case_ = WavaVecaConformerForPreTraining(UpperCAmelCase ) if is_finetuned: snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: snake_case_ = argparse.Namespace(task='audio_pretraining' ) snake_case_ = fairseq.tasks.setup_task(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=UpperCAmelCase ) snake_case_ = model[0].eval() recursively_load_weights(UpperCAmelCase , UpperCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __UpperCamelCase = parser.parse_args() convert_wavaveca_conformer_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
69
0
import os def A_ ( _UpperCAmelCase = "matrix.txt" ): with open(os.path.join(os.path.dirname(_UpperCAmelCase ) , _UpperCAmelCase ) ) as in_file: SCREAMING_SNAKE_CASE_: Union[str, Any] = in_file.read() SCREAMING_SNAKE_CASE_: int = [[int(_UpperCAmelCase ) for cell in row.split("," )] for row in data.strip().splitlines()] SCREAMING_SNAKE_CASE_: Dict = [[0 for cell in row] for row in grid] SCREAMING_SNAKE_CASE_: str = len(grid[0] ) SCREAMING_SNAKE_CASE_: Optional[int] = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] SCREAMING_SNAKE_CASE_: str = grid[0][0] for i in range(1 , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: int = grid[0][i] + dp[0][i - 1] for i in range(1 , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[Any] = grid[i][0] + dp[i - 1][0] for i in range(1 , _UpperCAmelCase ): for j in range(1 , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Dict = grid[i][j] + min(dp[i - 1][j] , dp[i][j - 1] ) return dp[-1][-1] if __name__ == "__main__": print(f'''{solution() = }''')
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase ) -> list: if len(UpperCAmelCase ) <= 1: return [tuple(UpperCAmelCase )] snake_case_ = [] def generate(UpperCAmelCase , UpperCAmelCase ): snake_case_ = [0] * n res.append(tuple(UpperCAmelCase ) ) snake_case_ = 0 while i < n: if c[i] < i: if i % 2 == 0: snake_case_ , snake_case_ = arr[i], arr[0] else: snake_case_ , snake_case_ = arr[i], arr[c[i]] res.append(tuple(UpperCAmelCase ) ) c[i] += 1 snake_case_ = 0 else: snake_case_ = 0 i += 1 generate(len(UpperCAmelCase ) , UpperCAmelCase ) return res if __name__ == "__main__": __UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
69
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """facebook/data2vec-vision-base-ft""": ( """https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json""" ), } class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): __lowerCamelCase : Any ='data2vec-vision' def __init__( self : Optional[int] , __lowercase : Dict=768 , __lowercase : int=12 , __lowercase : List[str]=12 , __lowercase : List[Any]=3072 , __lowercase : Dict="gelu" , __lowercase : List[Any]=0.0 , __lowercase : List[Any]=0.0 , __lowercase : List[str]=0.02 , __lowercase : Any=1E-12 , __lowercase : Dict=224 , __lowercase : str=16 , __lowercase : Any=3 , __lowercase : List[str]=False , __lowercase : Tuple=False , __lowercase : Any=False , __lowercase : List[Any]=False , __lowercase : Any=0.1 , __lowercase : Union[str, Any]=0.1 , __lowercase : Optional[int]=True , __lowercase : Dict=[3, 5, 7, 11] , __lowercase : Any=[1, 2, 3, 6] , __lowercase : Optional[Any]=True , __lowercase : Dict=0.4 , __lowercase : List[Any]=256 , __lowercase : str=1 , __lowercase : Any=False , __lowercase : List[Any]=255 , **__lowercase : Optional[Any] , ): '''simple docstring''' super().__init__(**lowerCAmelCase__ ) __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = initializer_range __a = layer_norm_eps __a = image_size __a = patch_size __a = num_channels __a = use_mask_token __a = use_absolute_position_embeddings __a = use_relative_position_bias __a = use_shared_relative_position_bias __a = layer_scale_init_value __a = drop_path_rate __a = use_mean_pooling # decode head attributes (semantic segmentation) __a = out_indices __a = pool_scales # auxiliary head attributes (semantic segmentation) __a = use_auxiliary_head __a = auxiliary_loss_weight __a = auxiliary_channels __a = auxiliary_num_convs __a = auxiliary_concat_input __a = semantic_loss_ignore_index class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): __lowerCamelCase : int =version.parse('1.11' ) @property def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' return 1E-4
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __UpperCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __UpperCamelCase = TaTokenizerFast __UpperCamelCase = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __UpperCamelCase = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
69
0
'''simple docstring''' from __future__ import annotations import math def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : int ) -> list[int]: if num <= 0: _a : List[Any] =F"{num}: Invalid input, please enter a positive integer." raise ValueError(_UpperCAmelCase ) _a : Union[str, Any] =[True] * (num + 1) _a : List[Any] =[] _a : List[str] =2 _a : List[str] =int(math.sqrt(_UpperCAmelCase ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(_UpperCAmelCase ) # Set multiples of start be False for i in range(start * start ,num + 1 ,_UpperCAmelCase ): if sieve[i] is True: _a : str =False start += 1 for j in range(end + 1 ,num + 1 ): if sieve[j] is True: prime.append(_UpperCAmelCase ) return prime if __name__ == "__main__": print(prime_sieve(int(input('''Enter a positive integer: ''').strip())))
276
"""simple docstring""" import sys from collections import defaultdict class UpperCamelCase : def __init__( self) -> Optional[int]: snake_case_ = [] def a_ ( self, lowerCAmelCase__) -> Any: return self.node_position[vertex] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = pos def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> str: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: snake_case_ = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: snake_case_ = 2 * start + 1 else: snake_case_ = 2 * start + 2 if heap[smallest_child] < heap[start]: snake_case_ , snake_case_ = heap[smallest_child], positions[smallest_child] snake_case_ , snake_case_ = ( heap[start], positions[start], ) snake_case_ , snake_case_ = temp, tempa snake_case_ = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child], self.get_position(positions[start])) self.set_position(positions[start], lowerCAmelCase__) self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> List[str]: snake_case_ = position[index] while index != 0: snake_case_ = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: snake_case_ = heap[parent] snake_case_ = position[parent] self.set_position(position[parent], lowerCAmelCase__) else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, lowerCAmelCase__) break snake_case_ = parent else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, 0) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = len(lowerCAmelCase__) // 2 - 1 for i in range(lowerCAmelCase__, -1, -1): self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, len(lowerCAmelCase__), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = positions[0] snake_case_ = sys.maxsize self.top_to_bottom(lowerCAmelCase__, 0, len(lowerCAmelCase__), lowerCAmelCase__) return temp def UpperCAmelCase ( UpperCAmelCase ) -> Tuple: snake_case_ = Heap() snake_case_ = [0] * len(UpperCAmelCase ) snake_case_ = [-1] * len(UpperCAmelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph snake_case_ = [] # Heap of Distance of vertices from their neighboring vertex snake_case_ = [] for vertex in range(len(UpperCAmelCase ) ): distance_tv.append(sys.maxsize ) positions.append(UpperCAmelCase ) heap.node_position.append(UpperCAmelCase ) snake_case_ = [] snake_case_ = 1 snake_case_ = sys.maxsize for neighbor, distance in adjacency_list[0]: snake_case_ = 0 snake_case_ = distance heap.heapify(UpperCAmelCase , UpperCAmelCase ) for _ in range(1 , len(UpperCAmelCase ) ): snake_case_ = heap.delete_minimum(UpperCAmelCase , UpperCAmelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) snake_case_ = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(UpperCAmelCase )] ): snake_case_ = distance heap.bottom_to_top( UpperCAmelCase , heap.get_position(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase ) snake_case_ = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > __UpperCamelCase = int(input('''Enter number of edges: ''').strip()) __UpperCamelCase = defaultdict(list) for _ in range(edges_number): __UpperCamelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
69
0
'''simple docstring''' from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class a__ : def __init__( self , _UpperCamelCase , ): """simple docstring""" _lowercase : int = parent _lowercase : Optional[Any] = 13 _lowercase : Optional[Any] = 7 _lowercase : List[str] = 30 _lowercase : Tuple = self.seq_length + self.mem_len _lowercase : Optional[Any] = 15 _lowercase : List[Any] = True _lowercase : Tuple = True _lowercase : List[str] = 99 _lowercase : int = [10, 50, 80] _lowercase : Tuple = 32 _lowercase : Dict = 32 _lowercase : Optional[Any] = 4 _lowercase : int = 8 _lowercase : str = 128 _lowercase : str = 2 _lowercase : Dict = 2 _lowercase : Tuple = None _lowercase : Optional[int] = 1 _lowercase : str = 0 _lowercase : Optional[Any] = 3 _lowercase : str = self.vocab_size - 1 _lowercase : List[str] = 0.0_1 def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : Tuple = None if self.use_labels: _lowercase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : List[Any] = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def _lowerCamelCase ( self ): """simple docstring""" random.seed(self.seed ) tf.random.set_seed(self.seed ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : List[Any] = TFTransfoXLModel(lowerCAmelCase__ ) _lowercase , _lowercase : Any = model(lowerCAmelCase__ ).to_tuple() _lowercase : int = {"input_ids": input_ids_a, "mems": mems_a} _lowercase , _lowercase : Optional[Any] = model(lowerCAmelCase__ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : Union[str, Any] = TFTransfoXLLMHeadModel(lowerCAmelCase__ ) _lowercase , _lowercase : Dict = model(lowerCAmelCase__ ).to_tuple() _lowercase : int = {"input_ids": input_ids_a, "labels": lm_labels} _lowercase , _lowercase : int = model(lowerCAmelCase__ ).to_tuple() _lowercase , _lowercase : List[Any] = model([input_ids_a, mems_a] ).to_tuple() _lowercase : int = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} _lowercase , _lowercase : List[Any] = model(lowerCAmelCase__ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : Optional[Any] = TFTransfoXLForSequenceClassification(lowerCAmelCase__ ) _lowercase : str = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Optional[Any] = self.prepare_config_and_inputs() ((_lowercase) , (_lowercase) , (_lowercase) , (_lowercase)) : Optional[Any] = config_and_inputs _lowercase : Optional[Any] = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class a__ ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Any = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) _SCREAMING_SNAKE_CASE : int = () if is_tf_available() else () _SCREAMING_SNAKE_CASE : Any = ( { 'feature-extraction': TFTransfoXLModel, 'text-classification': TFTransfoXLForSequenceClassification, 'text-generation': TFTransfoXLLMHeadModel, 'zero-shot': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented _SCREAMING_SNAKE_CASE : List[str] = False _SCREAMING_SNAKE_CASE : Any = False _SCREAMING_SNAKE_CASE : Union[str, Any] = False _SCREAMING_SNAKE_CASE : str = False def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Dict = TFTransfoXLModelTester(self ) _lowercase : Any = ConfigTester(self , config_class=lowerCAmelCase__ , d_embed=37 ) def _lowerCamelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def _lowerCamelCase ( self ): """simple docstring""" self.model_tester.set_seed() _lowercase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*lowerCAmelCase__ ) def _lowerCamelCase ( self ): """simple docstring""" self.model_tester.set_seed() _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*lowerCAmelCase__ ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*lowerCAmelCase__ ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase , _lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : int = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: _lowercase : List[Any] = model_class(lowerCAmelCase__ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: _lowercase : Tuple = model.get_output_embeddings() assert isinstance(lowerCAmelCase__ , tf.keras.layers.Layer ) _lowercase : Optional[int] = model.get_bias() assert name is None else: _lowercase : str = model.get_output_embeddings() assert x is None _lowercase : List[str] = model.get_bias() assert name is None def _lowerCamelCase ( self ): """simple docstring""" pass @slow def _lowerCamelCase ( self ): """simple docstring""" for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Optional[Any] = TFTransfoXLModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @unittest.skip(reason="This model doesn\'t play well with fit() due to not returning a single loss." ) def _lowerCamelCase ( self ): """simple docstring""" pass @require_tf class a__ ( unittest.TestCase ): @unittest.skip("Skip test until #12651 is resolved." ) @slow def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Any = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103" ) # fmt: off _lowercase : Union[str, Any] = tf.convert_to_tensor([[33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off _lowercase : Dict = [33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0,33,1,1857,2,1,1009,4,1109,11739,4762,358,5,25,245,28,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> _lowercase : Union[str, Any] = model.generate(lowerCAmelCase__ , max_length=200 , do_sample=lowerCAmelCase__ ) self.assertListEqual(output_ids[0].numpy().tolist() , lowerCAmelCase__ )
250
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase = { '''configuration_mctct''': ['''MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MCTCTConfig'''], '''feature_extraction_mctct''': ['''MCTCTFeatureExtractor'''], '''processing_mctct''': ['''MCTCTProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MCTCTForCTC''', '''MCTCTModel''', '''MCTCTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
'''simple docstring''' import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def __UpperCAmelCase ( ): with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(a_ ): requests.request("GET", "https://huggingface.co" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("GET", "https://huggingface.co", timeout=1.0 ) @pytest.mark.integration def __UpperCAmelCase ( ): with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("GET", "https://huggingface.co" ) def __UpperCAmelCase ( ): with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(a_ ): http_head("https://huggingface.co" )
145
"""simple docstring""" from math import factorial def UpperCAmelCase ( UpperCAmelCase = 20 ) -> int: snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(UpperCAmelCase ) / (factorial(UpperCAmelCase ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: __UpperCamelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
69
0
def a__ ( snake_case , snake_case ): """simple docstring""" _validate_point(snake_case ) _validate_point(snake_case ) if len(snake_case ) != len(snake_case ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(a - b ) for a, b in zip(snake_case , snake_case ) ) ) def a__ ( snake_case ): """simple docstring""" if point: if isinstance(snake_case , snake_case ): for item in point: if not isinstance(snake_case , (int, float) ): __SCREAMING_SNAKE_CASE : Any = ( '''Expected a list of numbers as input, found ''' F'''{type(snake_case ).__name__}''' ) raise TypeError(snake_case ) else: __SCREAMING_SNAKE_CASE : Union[str, Any] = F'''Expected a list of numbers as input, found {type(snake_case ).__name__}''' raise TypeError(snake_case ) else: raise ValueError('''Missing an input''' ) def a__ ( snake_case , snake_case ): """simple docstring""" _validate_point(snake_case ) _validate_point(snake_case ) if len(snake_case ) != len(snake_case ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(x - y ) for x, y in zip(snake_case , snake_case ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
303
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
0
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ : def __init__( self , _a , _a=13 , _a=32 , _a=3 , _a=4 , _a=[10, 20, 30, 40] , _a=[2, 2, 3, 2] , _a=True , _a=True , _a=37 , _a="gelu" , _a=10 , _a=0.0_2 , _a=["stage2", "stage3", "stage4"] , _a=3 , _a=None , ): lowercase : Optional[Any] = parent lowercase : str = batch_size lowercase : Optional[Any] = image_size lowercase : Dict = num_channels lowercase : Any = num_stages lowercase : Optional[Any] = hidden_sizes lowercase : Any = depths lowercase : Union[str, Any] = is_training lowercase : List[str] = use_labels lowercase : int = intermediate_size lowercase : str = hidden_act lowercase : str = type_sequence_label_size lowercase : Optional[int] = initializer_range lowercase : Optional[Any] = out_features lowercase : Tuple = num_labels lowercase : str = scope lowercase : str = num_stages def __magic_name__ ( self ): lowercase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase : List[Any] = None if self.use_labels: lowercase : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase : Tuple = self.get_config() return config, pixel_values, labels def __magic_name__ ( self ): return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def __magic_name__ ( self ): return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=lowerCAmelCase__ , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=lowerCAmelCase__ , loss_ignore_index=255 , num_labels=self.num_labels , ) def __magic_name__ ( self , _a , _a , _a ): lowercase : int = UperNetForSemanticSegmentation(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowercase : Optional[int] = model(lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def __magic_name__ ( self ): lowercase : Any = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ) : List[Any] = config_and_inputs lowercase : List[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a__ ( lowerCAmelCase__, lowerCAmelCase__, unittest.TestCase ): __lowerCAmelCase = (UperNetForSemanticSegmentation,) if is_torch_available() else () __lowerCAmelCase = {"""image-segmentation""": UperNetForSemanticSegmentation} if is_torch_available() else {} __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def __magic_name__ ( self ): lowercase : str = UperNetModelTester(self ) lowercase : Dict = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=37 ) def __magic_name__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __magic_name__ ( self ): return def __magic_name__ ( self ): lowercase , lowercase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : int = model_class(lowerCAmelCase__ ) lowercase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase : str = [*signature.parameters.keys()] lowercase : Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def __magic_name__ ( self ): lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCAmelCase__ ) @unittest.skip(reason="UperNet does not use inputs_embeds" ) def __magic_name__ ( self ): pass @unittest.skip(reason="UperNet does not support input and output embeddings" ) def __magic_name__ ( self ): pass @unittest.skip(reason="UperNet does not have a base model" ) def __magic_name__ ( self ): pass @unittest.skip(reason="UperNet does not have a base model" ) def __magic_name__ ( self ): pass @require_torch_multi_gpu @unittest.skip(reason="UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`" ) def __magic_name__ ( self ): pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __magic_name__ ( self ): pass def __magic_name__ ( self ): def check_hidden_states_output(_a , _a , _a ): lowercase : List[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): lowercase : Tuple = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) lowercase : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase : Optional[int] = self.model_tester.num_stages self.assertEqual(len(lowerCAmelCase__ ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase , lowercase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Optional[Any] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase : int = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __magic_name__ ( self ): lowercase , lowercase : Any = self.model_tester.prepare_config_and_inputs_for_common() lowercase : List[str] = _config_zero_init(lowerCAmelCase__ ) lowercase : Tuple = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: lowercase : List[str] = model_class(config=lowerCAmelCase__ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="UperNet does not have tied weights" ) def __magic_name__ ( self ): pass @slow def __magic_name__ ( self ): for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase : Optional[Any] = UperNetForSemanticSegmentation.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __magic_name__ ( ) -> str: lowercase : Optional[Any] = hf_hub_download( repo_id="hf-internal-testing/fixtures_ade20k" , repo_type="dataset" , filename="ADE_val_00000001.jpg" ) lowercase : List[Any] = Image.open(__snake_case ).convert("RGB" ) return image @require_torch @require_vision @slow class a__ ( unittest.TestCase ): def __magic_name__ ( self ): lowercase : Union[str, Any] = AutoImageProcessor.from_pretrained("openmmlab/upernet-swin-tiny" ) lowercase : str = UperNetForSemanticSegmentation.from_pretrained("openmmlab/upernet-swin-tiny" ).to(lowerCAmelCase__ ) lowercase : Optional[int] = prepare_img() lowercase : List[Any] = processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) with torch.no_grad(): lowercase : Any = model(**lowerCAmelCase__ ) lowercase : str = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) lowercase : str = torch.tensor( [[-7.5_9_5_8, -7.5_9_5_8, -7.4_3_0_2], [-7.5_9_5_8, -7.5_9_5_8, -7.4_3_0_2], [-7.4_7_9_7, -7.4_7_9_7, -7.3_0_6_8]] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) ) def __magic_name__ ( self ): lowercase : Union[str, Any] = AutoImageProcessor.from_pretrained("openmmlab/upernet-convnext-tiny" ) lowercase : Optional[Any] = UperNetForSemanticSegmentation.from_pretrained("openmmlab/upernet-convnext-tiny" ).to(lowerCAmelCase__ ) lowercase : List[Any] = prepare_img() lowercase : Optional[Any] = processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) with torch.no_grad(): lowercase : List[str] = model(**lowerCAmelCase__ ) lowercase : Tuple = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) lowercase : Tuple = torch.tensor( [[-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.7_7_4_6, -8.7_7_4_6, -8.6_1_3_0]] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) )
202
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = (DPMSolverSinglestepScheduler,) SCREAMING_SNAKE_CASE_ = (("num_inference_steps", 2_5),) def a_ ( self, **lowerCAmelCase__) -> int: snake_case_ = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf'), 'variance_type': None, } config.update(**lowerCAmelCase__) return config def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> List[Any]: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ , snake_case_ = sample, sample for t in range(lowerCAmelCase__, time_step + scheduler.config.solver_order + 1): snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self) -> Union[str, Any]: pass def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> int: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self, lowerCAmelCase__=None, **lowerCAmelCase__) -> Union[str, Any]: if scheduler is None: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample return sample def a_ ( self) -> List[Any]: snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = 50 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2574) < 1e-3 def a_ ( self) -> Dict: for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__) def a_ ( self) -> Optional[Any]: # make sure that iterating over schedulers with same config names gives same results # for defaults snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 snake_case_ = DEISMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverMultistepScheduler.from_config(scheduler.config) snake_case_ = UniPCMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverSinglestepScheduler.from_config(scheduler.config) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> str: self.check_over_configs(thresholding=lowerCAmelCase__) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCAmelCase__, prediction_type=lowerCAmelCase__, sample_max_value=lowerCAmelCase__, algorithm_type='dpmsolver++', solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, ) def a_ ( self) -> Tuple: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__) def a_ ( self) -> Optional[int]: for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) snake_case_ = self.full_loop( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) assert not torch.isnan(lowerCAmelCase__).any(), "Samples have nan numbers" def a_ ( self) -> Optional[Any]: self.check_over_configs(lower_order_final=lowerCAmelCase__) self.check_over_configs(lower_order_final=lowerCAmelCase__) def a_ ( self) -> Any: self.check_over_configs(lambda_min_clipped=-float('inf')) self.check_over_configs(lambda_min_clipped=-5.1) def a_ ( self) -> Any: self.check_over_configs(variance_type=lowerCAmelCase__) self.check_over_configs(variance_type='learned_range') def a_ ( self) -> List[Any]: for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=lowerCAmelCase__, time_step=0) def a_ ( self) -> int: snake_case_ = self.full_loop() snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> Dict: snake_case_ = self.full_loop(use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2248) < 1e-3 def a_ ( self) -> Union[str, Any]: snake_case_ = self.full_loop(prediction_type='v_prediction') snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.1453) < 1e-3 def a_ ( self) -> Optional[Any]: snake_case_ = self.full_loop(prediction_type='v_prediction', use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.0649) < 1e-3 def a_ ( self) -> Optional[int]: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(thresholding=lowerCAmelCase__, dynamic_thresholding_ratio=0) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample assert sample.dtype == torch.floataa
69
0
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class A__ ( lowerCAmelCase__ ): """simple docstring""" @require_torch def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : int = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' a__ : Any = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' a__ : Dict = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache a__ : List[str] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase__) BertModel.from_pretrained(lowerCAmelCase__) BertTokenizer.from_pretrained(lowerCAmelCase__) pipeline(task='fill-mask' , model=lowerCAmelCase__) # baseline - just load from_pretrained with normal network a__ : Union[str, Any] = [sys.executable, '-c', '\n'.join([load, run, mock])] # should succeed a__ : Optional[int] = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files a__ : List[str] = '1' a__ : List[str] = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 0 , result.stderr) self.assertIn('success' , result.stdout.decode()) @require_torch def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : Optional[Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' a__ : Union[str, Any] = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' a__ : Dict = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache a__ : List[str] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase__) BertModel.from_pretrained(lowerCAmelCase__) BertTokenizer.from_pretrained(lowerCAmelCase__) pipeline(task='fill-mask' , model=lowerCAmelCase__) # baseline - just load from_pretrained with normal network a__ : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock])] # should succeed a__ : Union[str, Any] = self.get_env() a__ : List[str] = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 0 , result.stderr) self.assertIn('success' , result.stdout.decode()) @require_torch def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : Optional[Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' a__ : Union[str, Any] = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' a__ : List[str] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network a__ : str = [sys.executable, '-c', '\n'.join([load, run])] # should succeed a__ : Optional[int] = self.get_env() a__ : Any = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 0 , result.stderr) self.assertIn('success' , result.stdout.decode()) # next emulate no network a__ : List[Any] = [sys.executable, '-c', '\n'.join([load, mock, run])] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files a__ : Tuple = '1' a__ : Optional[Any] = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 0 , result.stderr) self.assertIn('success' , result.stdout.decode()) @require_torch def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : str = '\nfrom transformers import pipeline\n ' a__ : Any = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' a__ : Any = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' a__ : Dict = self.get_env() a__ : Union[str, Any] = '1' a__ : Any = [sys.executable, '-c', '\n'.join([load, mock, run])] a__ : Union[str, Any] = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 1 , result.stderr) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '') , ) @require_torch def __lowercase ( self) -> List[str]: '''simple docstring''' a__ : List[Any] = '\nfrom transformers import AutoModel\n ' a__ : int = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network a__ : Optional[Any] = [sys.executable, '-c', '\n'.join([load, run])] # should succeed a__ : Optional[Any] = self.get_env() a__ : int = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 0 , result.stderr) self.assertIn('success' , result.stdout.decode()) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files a__ : Any = '1' a__ : List[str] = subprocess.run(lowerCAmelCase__ , env=lowerCAmelCase__ , check=lowerCAmelCase__ , capture_output=lowerCAmelCase__) self.assertEqual(result.returncode , 0 , result.stderr) self.assertIn('success' , result.stdout.decode())
99
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # Base Case if curr_ind == len(UpperCAmelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(UpperCAmelCase ) ): if valid_connection(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # Insert current vertex into path as next transition snake_case_ = next_ver # Validate created path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , curr_ind + 1 ): return True # Backtrack snake_case_ = -1 return False def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = 0 ) -> list[int]: snake_case_ = [-1] * (len(UpperCAmelCase ) + 1) # initialize start and end of path with starting index snake_case_ = snake_case_ = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , 1 ) else []
69
0
import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor A_ : int = logging.get_logger(__name__) class A_ ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , *lowercase__ , **lowercase__ ) -> None: warnings.warn( '''The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use VideoMAEImageProcessor instead.''' , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
333
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''▁''' __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model''' ), } } __UpperCamelCase = { '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__ = None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = {} if sp_model_kwargs is None else sp_model_kwargs snake_case_ = legacy_behaviour super().__init__( bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, sp_model_kwargs=self.sp_model_kwargs, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowerCAmelCase__)) snake_case_ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token snake_case_ = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab snake_case_ = 1 snake_case_ = len(self.sp_model) snake_case_ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase__) } snake_case_ = {v: k for k, v in self.lang_code_to_id.items()} snake_case_ = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) snake_case_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} snake_case_ = list(self.lang_code_to_id.keys()) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens]) snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.lang_code_to_id[self._src_lang] snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self) -> Union[str, Any]: snake_case_ = self.__dict__.copy() snake_case_ = None snake_case_ = self.sp_model.serialized_model_proto() return state def __setstate__( self, lowerCAmelCase__) -> Tuple: snake_case_ = d # for backward compatibility if not hasattr(self, 'sp_model_kwargs'): snake_case_ = {} snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) @property def a_ ( self) -> str: return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None, lowerCAmelCase__ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__, token_ids_a=lowerCAmelCase__, already_has_special_tokens=lowerCAmelCase__) snake_case_ = [1] * len(self.prefix_tokens) snake_case_ = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__)) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__)) + ([0] * len(lowerCAmelCase__)) + suffix_ones def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self) -> List[Any]: snake_case_ = {self.convert_ids_to_tokens(lowerCAmelCase__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def a_ ( self, lowerCAmelCase__) -> List[str]: return self.sp_model.encode(lowerCAmelCase__, out_type=lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> Any: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case_ = self.sp_model.PieceToId(lowerCAmelCase__) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a_ ( self, lowerCAmelCase__) -> Dict: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def a_ ( self, lowerCAmelCase__) -> List[str]: snake_case_ = ''.join(lowerCAmelCase__).replace(lowerCAmelCase__, ' ').strip() return out_string def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file, lowerCAmelCase__) elif not os.path.isfile(self.vocab_file): with open(lowerCAmelCase__, 'wb') as fi: snake_case_ = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__) return (out_vocab_file,) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> Union[str, Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> int: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[src_lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id]
69
0
import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Union[str, Any] = OmegaConf.load(__lowerCamelCase ) __snake_case : Any = torch.load(__lowerCamelCase , map_location="cpu" )["model"] __snake_case : Optional[int] = list(state_dict.keys() ) # extract state_dict for VQVAE __snake_case : List[Any] = {} __snake_case : Tuple = "first_stage_model." for key in keys: if key.startswith(__lowerCamelCase ): __snake_case : List[str] = state_dict[key] # extract state_dict for UNetLDM __snake_case : Tuple = {} __snake_case : int = "model.diffusion_model." for key in keys: if key.startswith(__lowerCamelCase ): __snake_case : Union[str, Any] = state_dict[key] __snake_case : Union[str, Any] = config.model.params.first_stage_config.params __snake_case : Tuple = config.model.params.unet_config.params __snake_case : Dict = VQModel(**__lowerCamelCase ).eval() vqvae.load_state_dict(__lowerCamelCase ) __snake_case : Any = UNetLDMModel(**__lowerCamelCase ).eval() unet.load_state_dict(__lowerCamelCase ) __snake_case : Dict = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=__lowerCamelCase , ) __snake_case : Union[str, Any] = LDMPipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) pipeline.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[str] = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", type=str, required=True) parser.add_argument("--config_path", type=str, required=True) parser.add_argument("--output_path", type=str, required=True) _snake_case : Optional[Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
123
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCAmelCase ( ) -> int: snake_case_ = HfArgumentParser(UpperCAmelCase ) snake_case_ = parser.parse_args_into_dataclasses()[0] snake_case_ = TensorFlowBenchmark(args=UpperCAmelCase ) try: snake_case_ = parser.parse_args_into_dataclasses()[0] except ValueError as e: snake_case_ = 'Arg --no_{0} is no longer used, please use --no-{0} instead.' snake_case_ = ' '.join(str(UpperCAmelCase ).split(' ' )[:-1] ) snake_case_ = '' snake_case_ = eval(str(UpperCAmelCase ).split(' ' )[-1] ) snake_case_ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = full_error_msg + begin_error_msg + str(UpperCAmelCase ) raise ValueError(UpperCAmelCase ) benchmark.run() if __name__ == "__main__": main()
69
0
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration SCREAMING_SNAKE_CASE__ : Dict = 50000 SCREAMING_SNAKE_CASE__ : int = 5000 SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = os.path.split(__file__) SCREAMING_SNAKE_CASE__ : int = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> str: for i in range(_SCREAMING_SNAKE_CASE ): lowerCamelCase : List[str] = dataset[i] @get_duration def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Tuple: for i in range(0 ,len(_SCREAMING_SNAKE_CASE ) ,_SCREAMING_SNAKE_CASE ): lowerCamelCase : Optional[Any] = dataset[i : i + batch_size] @get_duration def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[Any]: with dataset.formatted_as(type=_SCREAMING_SNAKE_CASE ): for i in range(_SCREAMING_SNAKE_CASE ): lowerCamelCase : Any = dataset[i] @get_duration def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: with dataset.formatted_as(type=_SCREAMING_SNAKE_CASE ): for i in range(0 ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): lowerCamelCase : List[str] = dataset[i : i + batch_size] def A ( ) -> Optional[Any]: lowerCamelCase : Optional[Any] = {"num examples": SPEED_TEST_N_EXAMPLES} lowerCamelCase : Dict = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] lowerCamelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) lowerCamelCase : Tuple = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) lowerCamelCase : List[Any] = generate_example_dataset( os.path.join(_SCREAMING_SNAKE_CASE ,"dataset.arrow" ) ,_SCREAMING_SNAKE_CASE ,num_examples=_SCREAMING_SNAKE_CASE ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(_SCREAMING_SNAKE_CASE ) ) lowerCamelCase : str = func(_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ) print("shuffling dataset" ) lowerCamelCase : Optional[Any] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(_SCREAMING_SNAKE_CASE ) ) lowerCamelCase : int = func( _SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE ,"wb" ) as f: f.write(json.dumps(_SCREAMING_SNAKE_CASE ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
48
"""simple docstring""" from __future__ import annotations def UpperCAmelCase ( UpperCAmelCase ) -> None: create_state_space_tree(UpperCAmelCase , [] , 0 , [0 for i in range(len(UpperCAmelCase ) )] ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) -> None: if index == len(UpperCAmelCase ): print(UpperCAmelCase ) return for i in range(len(UpperCAmelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) snake_case_ = True create_state_space_tree(UpperCAmelCase , UpperCAmelCase , index + 1 , UpperCAmelCase ) current_sequence.pop() snake_case_ = False __UpperCamelCase = [3, 1, 2, 4] generate_all_permutations(sequence) __UpperCamelCase = ["A", "B", "C"] generate_all_permutations(sequence_a)
69
0
import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Optional[int] = nn.functional.normalize(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] = nn.functional.normalize(_UpperCAmelCase ) return torch.mm(_UpperCAmelCase , normalized_text_embeds.t() ) class __lowercase ( lowerCAmelCase__ ): """simple docstring""" _UpperCAmelCase : str = CLIPConfig _UpperCAmelCase : Tuple = ['''CLIPEncoderLayer'''] def __init__( self : Tuple , lowerCAmelCase__ : Dict): super().__init__(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = CLIPVisionModel(config.vision_config) SCREAMING_SNAKE_CASE_: str = nn.Linear(config.vision_config.hidden_size , config.projection_dim , bias=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = nn.Parameter(torch.ones(17 , config.projection_dim) , requires_grad=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = nn.Parameter(torch.ones(3 , config.projection_dim) , requires_grad=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = nn.Parameter(torch.ones(17) , requires_grad=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = nn.Parameter(torch.ones(3) , requires_grad=lowerCAmelCase__) @torch.no_grad() def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : str): SCREAMING_SNAKE_CASE_: Union[str, Any] = self.vision_model(lowerCAmelCase__)[1] # pooled_output SCREAMING_SNAKE_CASE_: Optional[int] = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 SCREAMING_SNAKE_CASE_: Union[str, Any] = cosine_distance(lowerCAmelCase__ , self.special_care_embeds).cpu().float().numpy() SCREAMING_SNAKE_CASE_: Union[str, Any] = cosine_distance(lowerCAmelCase__ , self.concept_embeds).cpu().float().numpy() SCREAMING_SNAKE_CASE_: List[Any] = [] SCREAMING_SNAKE_CASE_: Dict = image_embeds.shape[0] for i in range(lowerCAmelCase__): SCREAMING_SNAKE_CASE_: List[Any] = {"special_scores": {}, "special_care": [], "concept_scores": {}, "bad_concepts": []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images SCREAMING_SNAKE_CASE_: Dict = 0.0 for concept_idx in range(len(special_cos_dist[0])): SCREAMING_SNAKE_CASE_: Dict = special_cos_dist[i][concept_idx] SCREAMING_SNAKE_CASE_: str = self.special_care_embeds_weights[concept_idx].item() SCREAMING_SNAKE_CASE_: Any = round(concept_cos - concept_threshold + adjustment , 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img["special_scores"][concept_idx]}) SCREAMING_SNAKE_CASE_: str = 0.01 for concept_idx in range(len(cos_dist[0])): SCREAMING_SNAKE_CASE_: List[Any] = cos_dist[i][concept_idx] SCREAMING_SNAKE_CASE_: int = self.concept_embeds_weights[concept_idx].item() SCREAMING_SNAKE_CASE_: Dict = round(concept_cos - concept_threshold + adjustment , 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = [len(res["bad_concepts"]) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[Any]): SCREAMING_SNAKE_CASE_: int = self.vision_model(lowerCAmelCase__)[1] # pooled_output SCREAMING_SNAKE_CASE_: List[str] = self.visual_projection(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = cosine_distance(lowerCAmelCase__ , self.special_care_embeds) SCREAMING_SNAKE_CASE_: str = cosine_distance(lowerCAmelCase__ , self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images SCREAMING_SNAKE_CASE_: List[Any] = 0.0 SCREAMING_SNAKE_CASE_: Optional[Any] = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) SCREAMING_SNAKE_CASE_: str = torch.any(special_scores > 0 , dim=1) SCREAMING_SNAKE_CASE_: List[str] = special_care * 0.01 SCREAMING_SNAKE_CASE_: Tuple = special_adjustment.unsqueeze(1).expand(-1 , cos_dist.shape[1]) SCREAMING_SNAKE_CASE_: int = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) SCREAMING_SNAKE_CASE_: Optional[int] = torch.any(concept_scores > 0 , dim=1) return images, has_nsfw_concepts
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( 'Warning: upper bound of deterministic test is exceeded. ' 'Pass allow_probable=True to allow probabilistic test. ' 'A return value of True indicates a probable prime.' ) # array bounds provided by analysis snake_case_ = [ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] snake_case_ = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(UpperCAmelCase , 1 ): if n < _p: # then we have our last prime to check snake_case_ = primes[:idx] break snake_case_ , snake_case_ = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: snake_case_ = False for r in range(UpperCAmelCase ): snake_case_ = pow(UpperCAmelCase , d * 2**r , UpperCAmelCase ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): snake_case_ = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def UpperCAmelCase ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
69
0
import os def lowerCAmelCase__ ( ): """simple docstring""" with open(os.path.dirname(_SCREAMING_SNAKE_CASE ) + """/p022_names.txt""" ) as file: __a = str(file.readlines()[0] ) __a = names.replace("""\"""" , """""" ).split(""",""" ) names.sort() __a = 0 __a = 0 for i, name in enumerate(_SCREAMING_SNAKE_CASE ): for letter in name: name_score += ord(_SCREAMING_SNAKE_CASE ) - 64 total_score += (i + 1) * name_score __a = 0 return total_score if __name__ == "__main__": print(solution())
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __UpperCamelCase = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
69
0
'''simple docstring''' def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : List[str] ,_UpperCAmelCase : Tuple = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( """Warning: upper bound of deterministic test is exceeded. """ """Pass allow_probable=True to allow probabilistic test. """ """A return value of True indicates a probable prime.""" ) # array bounds provided by analysis _a : Union[str, Any] =[ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] _a : Any =[2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(_UpperCAmelCase ,1 ): if n < _p: # then we have our last prime to check _a : Tuple =primes[:idx] break _a , _a : Union[str, Any] =n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: _a : List[Any] =False for r in range(_UpperCAmelCase ): _a : Tuple =pow(_UpperCAmelCase ,d * 2**r ,_UpperCAmelCase ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): _a : Optional[Any] =True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def SCREAMING_SNAKE_CASE_ ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
276
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: __UpperCamelCase = None __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } __UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = NllbTokenizer SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True snake_case_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) snake_case_ = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__) for lang_code in FAIRSEQ_LANGUAGE_CODES } snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.convert_tokens_to_ids(self._src_lang) snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> Tuple: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__): copyfile(self.vocab_file, lowerCAmelCase__) return (out_vocab_file,)
69
0
'''simple docstring''' import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() _snake_case = { 'bart': ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'bert': ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-base-cased-finetuned-mrpc': ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'dpr': ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'gpt2': ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlnet': ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm': ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm-roberta': ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'transfo-xl': ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'openai-gpt': ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'roberta': ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'layoutlm': ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'roberta-large-mnli': ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'camembert': ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'flaubert': ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert': ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert-base-distilled-squad': ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert': ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert-visual-feature-encoder': ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'ctrl': ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'albert': ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 't5': ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'electra': ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'wav2vec2': ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def _A ( snake_case , snake_case , snake_case , snake_case , snake_case=False , snake_case=True ) -> int: if model_type not in MODEL_CLASSES: raise ValueError(F'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) _lowercase , _lowercase , _lowercase , _lowercase : Dict = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: _lowercase : int = cached_file(snake_case , snake_case , force_download=not use_cached_models ) _lowercase : Any = config_class.from_json_file(snake_case ) _lowercase : Union[str, Any] = True _lowercase : Any = True print(F'''Building TensorFlow model from configuration: {config}''' ) _lowercase : Optional[Any] = model_class(snake_case ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): _lowercase : List[str] = cached_file( snake_case , snake_case , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: _lowercase : Tuple = load_pytorch_checkpoint_in_tfa_model(snake_case , snake_case ) if compare_with_pt_model: _lowercase : Optional[Any] = tf_model(tf_model.dummy_inputs , training=snake_case ) # build the network _lowercase : Union[str, Any] = torch.load(snake_case , map_location="cpu" ) _lowercase : int = pt_model_class.from_pretrained( pretrained_model_name_or_path=snake_case , config=snake_case , state_dict=snake_case ) with torch.no_grad(): _lowercase : Any = pt_model(**pt_model.dummy_inputs ) _lowercase : List[Any] = pto[0].numpy() _lowercase : int = tfo[0].numpy() _lowercase : int = np.amax(np.abs(np_pt - np_tf ) ) print(F'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2E-2, F'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(F'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(snake_case , save_format="h5" ) def _A ( snake_case , snake_case , snake_case=None , snake_case=None , snake_case=False , snake_case=False , snake_case=False , snake_case=False , ) -> Union[str, Any]: if args_model_type is None: _lowercase : str = list(MODEL_CLASSES.keys() ) else: _lowercase : int = [args_model_type] for j, model_type in enumerate(snake_case , start=1 ): print("=" * 1_00 ) print(F''' Converting model type {j}/{len(snake_case )}: {model_type}''' ) print("=" * 1_00 ) if model_type not in MODEL_CLASSES: raise ValueError(F'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) _lowercase , _lowercase , _lowercase , _lowercase , _lowercase : int = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: _lowercase : Tuple = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: _lowercase : str = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(snake_case , snake_case ) , start=1 ): print("-" * 1_00 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(F''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue _lowercase : Optional[Any] = model_shortcut_name elif only_convert_finetuned_models: print(F''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( F''' Converting checkpoint {i}/{len(snake_case )}: {model_shortcut_name} - model_type {model_type}''' ) print("-" * 1_00 ) if config_shortcut_name in aws_config_map: _lowercase : int = cached_file(snake_case , snake_case , force_download=not use_cached_models ) else: _lowercase : str = config_shortcut_name if model_shortcut_name in aws_model_maps: _lowercase : Dict = cached_file(snake_case , snake_case , force_download=not use_cached_models ) else: _lowercase : str = model_shortcut_name if os.path.isfile(snake_case ): _lowercase : Optional[int] = "converted_model" convert_pt_checkpoint_to_tf( model_type=snake_case , pytorch_checkpoint_path=snake_case , config_file=snake_case , tf_dump_path=os.path.join(snake_case , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=snake_case , ) if remove_cached_files: os.remove(snake_case ) os.remove(snake_case ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_dump_path', default=None, type=str, required=True, help='Path to the output Tensorflow dump file.' ) parser.add_argument( '--model_type', default=None, type=str, help=( F'''Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ''' 'convert all the models from AWS.' ), ) parser.add_argument( '--pytorch_checkpoint_path', default=None, type=str, help=( 'Path to the PyTorch checkpoint path or shortcut name to download from AWS. ' 'If not given, will download and convert all the checkpoints from AWS.' ), ) parser.add_argument( '--config_file', default=None, type=str, help=( 'The config json file corresponding to the pre-trained model. \n' 'This specifies the model architecture. If not given and ' '--pytorch_checkpoint_path is not given or is a shortcut name ' 'use the configuration associated to the shortcut name on the AWS' ), ) parser.add_argument( '--compare_with_pt_model', action='store_true', help='Compare Tensorflow and PyTorch model predictions.' ) parser.add_argument( '--use_cached_models', action='store_true', help='Use cached models if possible instead of updating to latest checkpoint versions.', ) parser.add_argument( '--remove_cached_files', action='store_true', help='Remove pytorch models after conversion (save memory when converting in batches).', ) parser.add_argument('--only_convert_finetuned_models', action='store_true', help='Only convert finetuned models.') _snake_case = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
250
"""simple docstring""" from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''T5Config''' class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig
69
0
'''simple docstring''' import math class A__ : """simple docstring""" def _lowerCAmelCase ( self : List[str] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Union[str, Any] ) -> int: """simple docstring""" _UpperCAmelCase : str = 0.0 _UpperCAmelCase : List[Any] = 0.0 for i in range(len(lowerCAmelCase__ ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : int , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Dict ) -> list[list[int | float]]: """simple docstring""" for i in range(len(lowerCAmelCase__ ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def __UpperCAmelCase ( ): # Training Examples ( m, n ) _UpperCAmelCase : List[Any] = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) _UpperCAmelCase : Any = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training _UpperCAmelCase : Any = SelfOrganizingMap() _UpperCAmelCase : str = 3 _UpperCAmelCase : List[str] = 0.5 for _ in range(a_ ): for j in range(len(a_ ) ): # training sample _UpperCAmelCase : Tuple = training_samples[j] # Compute the winning vector _UpperCAmelCase : List[Any] = self_organizing_map.get_winner(a_, a_ ) # Update the winning vector _UpperCAmelCase : List[Any] = self_organizing_map.update(a_, a_, a_, a_ ) # classify test sample _UpperCAmelCase : Optional[int] = [0, 0, 0, 1] _UpperCAmelCase : Union[str, Any] = self_organizing_map.get_winner(a_, a_ ) # results print(f"""Clusters that the test sample belongs to : {winner}""" ) print(f"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
145
"""simple docstring""" import argparse __UpperCamelCase = '''docs/source/_static/js/custom.js''' def UpperCAmelCase ( UpperCAmelCase ) -> int: with open(UpperCAmelCase , encoding='utf-8' , newline='\n' ) as f: snake_case_ = f.readlines() snake_case_ = 0 # First let's put the right version while not lines[index].startswith('const stableVersion =' ): index += 1 snake_case_ = f'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith('const versionMapping = {' ): index += 1 # We go until the end while not lines[index].startswith('}' ): index += 1 # We add the new version at the end lines[index - 1] += f' "v{version}": "v{version}",\n' with open(UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') __UpperCamelCase = parser.parse_args() update_custom_js(args.version)
69
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer lowercase_ = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast lowercase_ = TaTokenizerFast lowercase_ = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys lowercase_ = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
303
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class UpperCamelCase : def __init__( self, lowerCAmelCase__) -> Optional[int]: snake_case_ = data snake_case_ = None class UpperCamelCase : def __init__( self) -> Dict: snake_case_ = None snake_case_ = None def __iter__( self) -> Iterator[Any]: snake_case_ = self.head while self.head: yield node.data snake_case_ = node.next if node == self.head: break def __len__( self) -> int: return sum(1 for _ in self) def __repr__( self) -> str: return "->".join(str(lowerCAmelCase__) for item in iter(self)) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(len(self), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(0, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> None: if index < 0 or index > len(self): raise IndexError('list index out of range.') snake_case_ = Node(lowerCAmelCase__) if self.head is None: snake_case_ = new_node # first node points itself snake_case_ = snake_case_ = new_node elif index == 0: # insert at head snake_case_ = self.head snake_case_ = snake_case_ = new_node else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = new_node if index == len(self) - 1: # insert at tail snake_case_ = new_node def a_ ( self) -> str: return self.delete_nth(0) def a_ ( self) -> Any: return self.delete_nth(len(self) - 1) def a_ ( self, lowerCAmelCase__ = 0) -> Any: if not 0 <= index < len(self): raise IndexError('list index out of range.') snake_case_ = self.head if self.head == self.tail: # just one node snake_case_ = snake_case_ = None elif index == 0: # delete head node snake_case_ = self.tail.next.next snake_case_ = self.head.next else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = temp.next.next if index == len(self) - 1: # delete at tail snake_case_ = temp return delete_node.data def a_ ( self) -> bool: return len(self) == 0 def UpperCAmelCase ( ) -> None: snake_case_ = CircularLinkedList() assert len(UpperCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(UpperCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(UpperCAmelCase ) == i circular_linked_list.insert_nth(UpperCAmelCase , i + 1 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
69
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__ ( lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, unittest.TestCase ): __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline __lowerCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} __lowerCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __lowerCAmelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS __lowerCAmelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def __magic_name__ ( self ): torch.manual_seed(0 ) lowercase : List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) lowercase : List[str] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) torch.manual_seed(0 ) lowercase : List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) lowercase : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) lowercase : List[str] = CLIPTextModel(lowerCAmelCase__ ) lowercase : Tuple = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowercase : Union[str, Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def __magic_name__ ( self , _a , _a=0 ): lowercase : List[str] = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowercase : int = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase : List[str] = Image.fromarray(np.uinta(lowerCAmelCase__ ) ).convert("RGB" ) if str(lowerCAmelCase__ ).startswith("mps" ): lowercase : int = torch.manual_seed(lowerCAmelCase__ ) else: lowercase : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowercase : Union[str, Any] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "image_guidance_scale": 1, "output_type": "numpy", } return inputs def __magic_name__ ( self ): lowercase : List[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase : Union[str, Any] = self.get_dummy_components() lowercase : Any = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase__ ) lowercase : Tuple = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowercase : Optional[Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowercase : Optional[Any] = sd_pipe(**lowerCAmelCase__ ).images lowercase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowercase : Dict = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__ ( self ): lowercase : Optional[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase : Union[str, Any] = self.get_dummy_components() lowercase : List[str] = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase__ ) lowercase : Optional[Any] = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowercase : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowercase : List[Any] = "french fries" lowercase : str = sd_pipe(**lowerCAmelCase__ , negative_prompt=lowerCAmelCase__ ) lowercase : Optional[Any] = output.images lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowercase : List[Any] = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__ ( self ): lowercase : List[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase : Tuple = self.get_dummy_components() lowercase : int = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase__ ) lowercase : Dict = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowercase : int = self.get_dummy_inputs(lowerCAmelCase__ ) lowercase : Union[str, Any] = [inputs["prompt"]] * 2 lowercase : Tuple = np.array(inputs["image"] ).astype(np.floataa ) / 2_5_5.0 lowercase : Any = torch.from_numpy(lowerCAmelCase__ ).unsqueeze(0 ).to(lowerCAmelCase__ ) lowercase : List[str] = image / 2 + 0.5 lowercase : Optional[Any] = image.permute(0 , 3 , 1 , 2 ) lowercase : str = image.repeat(2 , 1 , 1 , 1 ) lowercase : List[Any] = sd_pipe(**lowerCAmelCase__ ).images lowercase : Optional[int] = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) lowercase : Tuple = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__ ( self ): lowercase : List[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase : str = self.get_dummy_components() lowercase : int = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" ) lowercase : Optional[Any] = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase__ ) lowercase : Optional[Any] = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowercase : int = self.get_dummy_inputs(lowerCAmelCase__ ) lowercase : int = sd_pipe(**lowerCAmelCase__ ).images lowercase : Dict = image[0, -3:, -3:, -1] lowercase : List[str] = [round(lowerCAmelCase__ , 4 ) for x in image_slice.flatten().tolist()] print(",".join([str(lowerCAmelCase__ ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) lowercase : Tuple = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def __magic_name__ ( self ): lowercase : Any = self.get_dummy_components() lowercase : Tuple = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase__ ) lowercase : Dict = VaeImageProcessor(do_resize=lowerCAmelCase__ , do_normalize=lowerCAmelCase__ ) lowercase : int = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowercase : str = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase__ , input_image_type="pt" ) )[0] lowercase : str = components["vae"] lowercase : Any = self.get_dummy_inputs_by_type(lowerCAmelCase__ , input_image_type="pt" ) for image_param in self.image_latents_params: if image_param in inputs.keys(): lowercase : int = vae.encode(inputs[image_param] ).latent_dist.mode() lowercase : Dict = pipe(**lowerCAmelCase__ )[0] lowercase : Tuple = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase__ , 1E-4 , "passing latents as image input generate different result from passing image" ) @slow @require_torch_gpu class a__ ( unittest.TestCase ): def __magic_name__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ ( self , _a=0 ): lowercase : Tuple = torch.manual_seed(lowerCAmelCase__ ) lowercase : List[str] = load_image( "https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg" ) lowercase : Any = { "prompt": "turn him into a cyborg", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "image_guidance_scale": 1.0, "output_type": "numpy", } return inputs def __magic_name__ ( self ): lowercase : Optional[int] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=lowerCAmelCase__ ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() lowercase : Any = self.get_inputs() lowercase : Any = pipe(**lowerCAmelCase__ ).images lowercase : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowercase : int = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def __magic_name__ ( self ): lowercase : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=lowerCAmelCase__ ) lowercase : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() lowercase : List[Any] = self.get_inputs() lowercase : Any = pipe(**lowerCAmelCase__ ).images lowercase : List[str] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowercase : List[str] = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def __magic_name__ ( self ): lowercase : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=lowerCAmelCase__ ) lowercase : int = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() lowercase : Dict = self.get_inputs() lowercase : Dict = pipe(**lowerCAmelCase__ ).images lowercase : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowercase : Optional[int] = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def __magic_name__ ( self ): lowercase : Union[str, Any] = 0 def callback_fn(_a , _a , _a ) -> None: lowercase : Optional[int] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: lowercase : int = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) lowercase : Any = latents[0, -3:, -3:, -1] lowercase : int = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: lowercase : Any = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) lowercase : int = latents[0, -3:, -3:, -1] lowercase : Union[str, Any] = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 lowercase : Optional[int] = False lowercase : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=lowerCAmelCase__ , torch_dtype=torch.floataa ) lowercase : List[Any] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() lowercase : Union[str, Any] = self.get_inputs() pipe(**lowerCAmelCase__ , callback=lowerCAmelCase__ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def __magic_name__ ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowercase : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( "timbrooks/instruct-pix2pix" , safety_checker=lowerCAmelCase__ , torch_dtype=torch.floataa ) lowercase : List[Any] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowercase : List[Any] = self.get_inputs() lowercase : Any = pipe(**lowerCAmelCase__ ) lowercase : Tuple = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def __magic_name__ ( self ): lowercase : str = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 lowercase : List[Any] = inputs["image"].resize((504, 504) ) lowercase : List[Any] = "timbrooks/instruct-pix2pix" lowercase : int = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() lowercase : Optional[Any] = pipe(**lowerCAmelCase__ ) lowercase : Tuple = output.images[0] lowercase : Optional[int] = image[255:258, 383:386, -1] assert image.shape == (504, 504, 3) lowercase : Dict = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
202
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCamelCase = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __UpperCamelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) __UpperCamelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __UpperCamelCase = re.compile(r'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __UpperCamelCase = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def UpperCAmelCase ( UpperCAmelCase ) -> List[Any]: snake_case_ = None # source code of `config_class` snake_case_ = inspect.getsource(UpperCAmelCase ) snake_case_ = _re_checkpoint.findall(UpperCAmelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): snake_case_ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: snake_case_ = ckpt_name break return checkpoint def UpperCAmelCase ( ) -> Union[str, Any]: snake_case_ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ = get_checkpoint_from_config_class(UpperCAmelCase ) snake_case_ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = '\n'.join(sorted(UpperCAmelCase ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class A__ : """simple docstring""" __A : List[str] = BlenderbotConfig __A : Tuple = {} __A : Optional[int] = '''gelu''' def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=False , lowercase=99 , lowercase=32 , lowercase=2 , lowercase=4 , lowercase=37 , lowercase=0.1 , lowercase=0.1 , lowercase=20 , lowercase=2 , lowercase=1 , lowercase=0 , ) -> Optional[int]: '''simple docstring''' a__ : Union[str, Any] = parent a__ : Dict = batch_size a__ : Union[str, Any] = seq_length a__ : List[str] = is_training a__ : List[Any] = use_labels a__ : str = vocab_size a__ : str = hidden_size a__ : List[Any] = num_hidden_layers a__ : Tuple = num_attention_heads a__ : Dict = intermediate_size a__ : Any = hidden_dropout_prob a__ : List[Any] = attention_probs_dropout_prob a__ : List[str] = max_position_embeddings a__ : List[Any] = eos_token_id a__ : Union[str, Any] = pad_token_id a__ : str = bos_token_id def __lowercase ( self) -> List[str]: '''simple docstring''' a__ : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) a__ : str = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) a__ : List[str] = tf.concat([input_ids, eos_tensor] , axis=1) a__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a__ : Optional[Any] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) a__ : Optional[int] = prepare_blenderbot_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) return config, inputs_dict def __lowercase ( self , lowercase , lowercase) -> Dict: '''simple docstring''' a__ : int = TFBlenderbotModel(config=lowerCAmelCase__).get_decoder() a__ : Any = inputs_dict['input_ids'] a__ : List[str] = input_ids[:1, :] a__ : Any = inputs_dict['attention_mask'][:1, :] a__ : Dict = inputs_dict['head_mask'] a__ : List[str] = 1 # first forward pass a__ : Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , head_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__) a__ , a__ : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids a__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size) a__ : Dict = tf.cast(ids_tensor((self.batch_size, 3) , 2) , tf.inta) # append to next input_ids and a__ : List[Any] = tf.concat([input_ids, next_tokens] , axis=-1) a__ : Union[str, Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1) a__ : List[str] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__)[0] a__ : str = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__)[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1]) # select random slice a__ : int = int(ids_tensor((1,) , output_from_past.shape[-1])) a__ : Any = output_from_no_past[:, -3:, random_slice_idx] a__ : int = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , rtol=1e-3) def A_ ( A__ , A__ , A__ , A__=None , A__=None , A__=None , A__=None , A__=None , ) -> Optional[Any]: if attention_mask is None: a__ : int = tf.cast(tf.math.not_equal(A__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: a__ : List[str] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: a__ : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: a__ : Union[str, Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: a__ : List[str] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class A__ ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" __A : Dict = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __A : str = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __A : int = ( { '''conversational''': TFBlenderbotForConditionalGeneration, '''feature-extraction''': TFBlenderbotModel, '''summarization''': TFBlenderbotForConditionalGeneration, '''text2text-generation''': TFBlenderbotForConditionalGeneration, '''translation''': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __A : List[str] = True __A : List[str] = False __A : List[str] = False def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : int = TFBlenderbotModelTester(self) a__ : Optional[int] = ConfigTester(self , config_class=lowerCAmelCase__) def __lowercase ( self) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase__) @require_tokenizers @require_tf class A__ ( unittest.TestCase ): """simple docstring""" __A : int = ['''My friends are cool but they eat too many carbs.'''] __A : Any = '''facebook/blenderbot-400M-distill''' @cached_property def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' return BlenderbotTokenizer.from_pretrained(self.model_name) @cached_property def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model @slow def __lowercase ( self) -> Dict: '''simple docstring''' a__ : Dict = self.tokenizer(self.src_text , return_tensors='tf') a__ : List[Any] = self.model.generate( model_inputs.input_ids , ) a__ : Optional[int] = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=lowerCAmelCase__)[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
99
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __UpperCamelCase = 0 __UpperCamelCase = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __UpperCamelCase = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __UpperCamelCase = tuple[int, int] class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, ) -> None: snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = g_cost snake_case_ = parent snake_case_ = self.calculate_heuristic() snake_case_ = self.g_cost + self.h_cost def a_ ( self) -> float: snake_case_ = self.pos_x - self.goal_x snake_case_ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase__) + abs(lowerCAmelCase__) else: return sqrt(dy**2 + dx**2) def __lt__( self, lowerCAmelCase__) -> bool: return self.f_cost < other.f_cost class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = Node(start[1], start[0], goal[1], goal[0], 0, lowerCAmelCase__) snake_case_ = Node(goal[1], goal[0], goal[1], goal[0], 9_9999, lowerCAmelCase__) snake_case_ = [self.start] snake_case_ = [] snake_case_ = False def a_ ( self) -> list[TPosition]: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() snake_case_ = self.open_nodes.pop(0) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase__) self.closed_nodes.append(lowerCAmelCase__) snake_case_ = self.get_successors(lowerCAmelCase__) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__) else: self.open_nodes.append(lowerCAmelCase__) return [self.start.pos] def a_ ( self, lowerCAmelCase__) -> list[Node]: snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(lowerCAmelCase__) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__, lowerCAmelCase__, self.target.pos_y, self.target.pos_x, parent.g_cost + 1, lowerCAmelCase__, )) return successors def a_ ( self, lowerCAmelCase__) -> list[TPosition]: snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x)) snake_case_ = current_node.parent path.reverse() return path class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> None: snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = False def a_ ( self) -> list[TPosition]: while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() snake_case_ = self.fwd_astar.open_nodes.pop(0) snake_case_ = self.bwd_astar.open_nodes.pop(0) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase__, lowerCAmelCase__) self.fwd_astar.closed_nodes.append(lowerCAmelCase__) self.bwd_astar.closed_nodes.append(lowerCAmelCase__) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase__), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase__), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase__) else: astar.open_nodes.append(lowerCAmelCase__) return [self.fwd_astar.start.pos] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> list[TPosition]: snake_case_ = self.fwd_astar.retrace_path(lowerCAmelCase__) snake_case_ = self.bwd_astar.retrace_path(lowerCAmelCase__) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __UpperCamelCase = (0, 0) __UpperCamelCase = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __UpperCamelCase = time.time() __UpperCamelCase = AStar(init, goal) __UpperCamelCase = a_star.search() __UpperCamelCase = time.time() - start_time print(F"""AStar execution time = {end_time:f} seconds""") __UpperCamelCase = time.time() __UpperCamelCase = BidirectionalAStar(init, goal) __UpperCamelCase = time.time() - bd_start_time print(F"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
69
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ : List[Any] = { 'configuration_bert': ['BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BertConfig', 'BertOnnxConfig'], 'tokenization_bert': ['BasicTokenizer', 'BertTokenizer', 'WordpieceTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : int = ['BertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : int = [ 'BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BertForMaskedLM', 'BertForMultipleChoice', 'BertForNextSentencePrediction', 'BertForPreTraining', 'BertForQuestionAnswering', 'BertForSequenceClassification', 'BertForTokenClassification', 'BertLayer', 'BertLMHeadModel', 'BertModel', 'BertPreTrainedModel', 'load_tf_weights_in_bert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : str = [ 'TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFBertEmbeddings', 'TFBertForMaskedLM', 'TFBertForMultipleChoice', 'TFBertForNextSentencePrediction', 'TFBertForPreTraining', 'TFBertForQuestionAnswering', 'TFBertForSequenceClassification', 'TFBertForTokenClassification', 'TFBertLMHeadModel', 'TFBertMainLayer', 'TFBertModel', 'TFBertPreTrainedModel', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : int = ['TFBertTokenizer'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Dict = [ 'FlaxBertForCausalLM', 'FlaxBertForMaskedLM', 'FlaxBertForMultipleChoice', 'FlaxBertForNextSentencePrediction', 'FlaxBertForPreTraining', 'FlaxBertForQuestionAnswering', 'FlaxBertForSequenceClassification', 'FlaxBertForTokenClassification', 'FlaxBertModel', 'FlaxBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys A_ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
333
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: while a != 0: snake_case_ , snake_case_ = b % a, a return b def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: if gcd(UpperCAmelCase , UpperCAmelCase ) != 1: snake_case_ = f'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = 1, 0, a snake_case_ , snake_case_ , snake_case_ = 0, 1, m while va != 0: snake_case_ = ua // va snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
69
0
from __future__ import annotations class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> str: __snake_case , __snake_case : Dict = text, pattern __snake_case , __snake_case : Optional[int] = len(lowerCAmelCase__ ), len(lowerCAmelCase__ ) def __snake_case ( self : List[Any] , lowerCamelCase : Optional[Any] ) -> int: for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] ) -> int: for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def __snake_case ( self : List[Any] ) -> list[int]: # searches pattern in text and returns index positions __snake_case : Optional[int] = [] for i in range(self.textLen - self.patLen + 1 ): __snake_case : Optional[int] = self.mismatch_in_text(lowerCAmelCase__ ) if mismatch_index == -1: positions.append(lowerCAmelCase__ ) else: __snake_case : Tuple = self.match_in_pattern(self.text[mismatch_index] ) __snake_case : Optional[int] = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _snake_case : str = "ABAABA" _snake_case : Optional[int] = "AB" _snake_case : Any = BoyerMooreSearch(text, pattern) _snake_case : Optional[Any] = bms.bad_character_heuristic() if len(positions) == 0: print("No match found") else: print("Pattern found in following positions: ") print(positions)
123
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase = { '''configuration_bert''': ['''BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BertConfig''', '''BertOnnxConfig'''], '''tokenization_bert''': ['''BasicTokenizer''', '''BertTokenizer''', '''WordpieceTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''BertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BertForMaskedLM''', '''BertForMultipleChoice''', '''BertForNextSentencePrediction''', '''BertForPreTraining''', '''BertForQuestionAnswering''', '''BertForSequenceClassification''', '''BertForTokenClassification''', '''BertLayer''', '''BertLMHeadModel''', '''BertModel''', '''BertPreTrainedModel''', '''load_tf_weights_in_bert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBertEmbeddings''', '''TFBertForMaskedLM''', '''TFBertForMultipleChoice''', '''TFBertForNextSentencePrediction''', '''TFBertForPreTraining''', '''TFBertForQuestionAnswering''', '''TFBertForSequenceClassification''', '''TFBertForTokenClassification''', '''TFBertLMHeadModel''', '''TFBertMainLayer''', '''TFBertModel''', '''TFBertPreTrainedModel''', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFBertTokenizer'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxBertForCausalLM''', '''FlaxBertForMaskedLM''', '''FlaxBertForMultipleChoice''', '''FlaxBertForNextSentencePrediction''', '''FlaxBertForPreTraining''', '''FlaxBertForQuestionAnswering''', '''FlaxBertForSequenceClassification''', '''FlaxBertForTokenClassification''', '''FlaxBertModel''', '''FlaxBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
def A ( _SCREAMING_SNAKE_CASE ) -> bool: if not all(x.isalpha() for x in string ): raise ValueError("String must only contain alphabetic characters." ) lowerCamelCase : Dict = sorted(string.lower() ) return len(_SCREAMING_SNAKE_CASE ) == len(set(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Optional[int] = input('Enter a string ').strip() SCREAMING_SNAKE_CASE__ : Any = is_isogram(input_str) print(f'''{input_str} is {"an" if isogram else "not an"} isogram.''')
48
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.linear_k''': '''encoder.layers.*.self_attn.linear_k''', '''self_attn.linear_v''': '''encoder.layers.*.self_attn.linear_v''', '''self_attn.linear_q''': '''encoder.layers.*.self_attn.linear_q''', '''self_attn.pos_bias_u''': '''encoder.layers.*.self_attn.pos_bias_u''', '''self_attn.pos_bias_v''': '''encoder.layers.*.self_attn.pos_bias_v''', '''self_attn.linear_out''': '''encoder.layers.*.self_attn.linear_out''', '''self_attn.linear_pos''': '''encoder.layers.*.self_attn.linear_pos''', '''self_attn.rotary_emb''': '''encoder.embed_positions''', '''self_attn_layer_norm''': '''encoder.layers.*.self_attn_layer_norm''', '''conv_module.pointwise_conv1''': '''encoder.layers.*.conv_module.pointwise_conv1''', '''conv_module.pointwise_conv2''': '''encoder.layers.*.conv_module.pointwise_conv2''', '''conv_module.depthwise_conv''': '''encoder.layers.*.conv_module.depthwise_conv''', '''conv_module.batch_norm''': '''encoder.layers.*.conv_module.batch_norm''', '''conv_module.layer_norm''': '''encoder.layers.*.conv_module.layer_norm''', '''ffn1.w_1''': '''encoder.layers.*.ffn1.intermediate_dense''', '''ffn1.w_2''': '''encoder.layers.*.ffn1.output_dense''', '''ffn1.layer_norm''': '''encoder.layers.*.ffn1_layer_norm''', '''ffn2.w_1''': '''encoder.layers.*.ffn2.intermediate_dense''', '''ffn2.w_2''': '''encoder.layers.*.ffn2.output_dense''', '''ffn2.layer_norm''': '''encoder.layers.*.ffn2_layer_norm''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __UpperCamelCase = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: for attribute in key.split('.' ): snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ) if weight_type is not None: snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ).shape else: snake_case_ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": snake_case_ = value elif weight_type == "weight_g": snake_case_ = value elif weight_type == "weight_v": snake_case_ = value elif weight_type == "bias": snake_case_ = value elif weight_type == "running_mean": snake_case_ = value elif weight_type == "running_var": snake_case_ = value elif weight_type == "num_batches_tracked": snake_case_ = value elif weight_type == "inv_freq": snake_case_ = value else: snake_case_ = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = [] snake_case_ = fairseq_model.state_dict() snake_case_ = hf_model.wavaveca_conformer.feature_extractor for name, value in fairseq_dict.items(): snake_case_ = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) snake_case_ = True else: for key, mapped_key in MAPPING.items(): snake_case_ = 'wav2vec2_conformer.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case_ = True if "*" in mapped_key: snake_case_ = name.split(UpperCAmelCase )[0].split('.' )[-2] snake_case_ = mapped_key.replace('*' , UpperCAmelCase ) if "pos_bias_u" in name: snake_case_ = None elif "pos_bias_v" in name: snake_case_ = None elif "weight_g" in name: snake_case_ = 'weight_g' elif "weight_v" in name: snake_case_ = 'weight_v' elif "bias" in name: snake_case_ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case_ = 'weight' elif "running_mean" in name: snake_case_ = 'running_mean' elif "inv_freq" in name: snake_case_ = 'inv_freq' elif "running_var" in name: snake_case_ = 'running_var' elif "num_batches_tracked" in name: snake_case_ = 'num_batches_tracked' else: snake_case_ = None set_recursively(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) continue if not is_used: unused_weights.append(UpperCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: snake_case_ = full_name.split('conv_layers.' )[-1] snake_case_ = name.split('.' ) snake_case_ = int(items[0] ) snake_case_ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase ) @torch.no_grad() def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=True ) -> str: if config_path is not None: snake_case_ = WavaVecaConformerConfig.from_pretrained(UpperCAmelCase , hidden_act='swish' ) else: snake_case_ = WavaVecaConformerConfig() if "rope" in checkpoint_path: snake_case_ = 'rotary' if is_finetuned: if dict_path: snake_case_ = Dictionary.load(UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq snake_case_ = target_dict.pad_index snake_case_ = target_dict.bos_index snake_case_ = target_dict.eos_index snake_case_ = len(target_dict.symbols ) snake_case_ = os.path.join(UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase ) ) return os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) snake_case_ = target_dict.indices # fairseq has the <pad> and <s> switched snake_case_ = 0 snake_case_ = 1 with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(UpperCAmelCase , UpperCAmelCase ) snake_case_ = WavaVecaCTCTokenizer( UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase , ) snake_case_ = True if config.feat_extract_norm == 'layer' else False snake_case_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=UpperCAmelCase , return_attention_mask=UpperCAmelCase , ) snake_case_ = WavaVecaProcessor(feature_extractor=UpperCAmelCase , tokenizer=UpperCAmelCase ) processor.save_pretrained(UpperCAmelCase ) snake_case_ = WavaVecaConformerForCTC(UpperCAmelCase ) else: snake_case_ = WavaVecaConformerForPreTraining(UpperCAmelCase ) if is_finetuned: snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: snake_case_ = argparse.Namespace(task='audio_pretraining' ) snake_case_ = fairseq.tasks.setup_task(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=UpperCAmelCase ) snake_case_ = model[0].eval() recursively_load_weights(UpperCAmelCase , UpperCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __UpperCamelCase = parser.parse_args() convert_wavaveca_conformer_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
69
0
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class __lowercase ( lowerCAmelCase__ ): """simple docstring""" _UpperCAmelCase : Optional[int] = 42 _UpperCAmelCase : Optional[int] = None def A_ ( _UpperCAmelCase , _UpperCAmelCase=0.9_9_9 , _UpperCAmelCase="cosine" , ): if alpha_transform_type == "cosine": def alpha_bar_fn(_UpperCAmelCase ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_UpperCAmelCase ): return math.exp(t * -1_2.0 ) else: raise ValueError(f"Unsupported alpha_tranform_type: {alpha_transform_type}" ) SCREAMING_SNAKE_CASE_: Any = [] for i in range(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Union[str, Any] = i / num_diffusion_timesteps SCREAMING_SNAKE_CASE_: str = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(_UpperCAmelCase ) / alpha_bar_fn(_UpperCAmelCase ) , _UpperCAmelCase ) ) return torch.tensor(_UpperCAmelCase , dtype=torch.floataa ) class __lowercase ( lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" @register_to_config def __init__( self : str , lowerCAmelCase__ : Union[str, Any] = 1000 , lowerCAmelCase__ : Union[str, Any] = "fixed_small_log" , lowerCAmelCase__ : Any = True , lowerCAmelCase__ : Any = 1.0 , lowerCAmelCase__ : Union[str, Any] = "epsilon" , lowerCAmelCase__ : Optional[Any] = "squaredcos_cap_v2" , ): if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: \'squaredcos_cap_v2\'") SCREAMING_SNAKE_CASE_: List[Any] = betas_for_alpha_bar(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Tuple = 1.0 - self.betas SCREAMING_SNAKE_CASE_: Union[str, Any] = torch.cumprod(self.alphas , dim=0) SCREAMING_SNAKE_CASE_: Tuple = torch.tensor(1.0) # standard deviation of the initial noise distribution SCREAMING_SNAKE_CASE_: Any = 1.0 # setable values SCREAMING_SNAKE_CASE_: str = None SCREAMING_SNAKE_CASE_: List[str] = torch.from_numpy(np.arange(0 , lowerCAmelCase__)[::-1].copy()) SCREAMING_SNAKE_CASE_: Optional[int] = variance_type def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase__ : Any , lowerCAmelCase__ : Any = None): return sample def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Tuple = None): SCREAMING_SNAKE_CASE_: Any = num_inference_steps SCREAMING_SNAKE_CASE_: str = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) SCREAMING_SNAKE_CASE_: int = (np.arange(0 , lowerCAmelCase__) * step_ratio).round()[::-1].copy().astype(np.intaa) SCREAMING_SNAKE_CASE_: int = torch.from_numpy(lowerCAmelCase__).to(lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Tuple=None , lowerCAmelCase__ : Optional[Any]=None , lowerCAmelCase__ : str=None): if prev_timestep is None: SCREAMING_SNAKE_CASE_: Union[str, Any] = t - 1 SCREAMING_SNAKE_CASE_: List[str] = self.alphas_cumprod[t] SCREAMING_SNAKE_CASE_: Optional[int] = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one SCREAMING_SNAKE_CASE_: Optional[int] = 1 - alpha_prod_t SCREAMING_SNAKE_CASE_: List[Any] = 1 - alpha_prod_t_prev if prev_timestep == t - 1: SCREAMING_SNAKE_CASE_: str = self.betas[t] else: SCREAMING_SNAKE_CASE_: Optional[int] = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample SCREAMING_SNAKE_CASE_: Union[str, Any] = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: SCREAMING_SNAKE_CASE_: List[Any] = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": SCREAMING_SNAKE_CASE_: Optional[int] = torch.log(torch.clamp(lowerCAmelCase__ , min=1E-20)) SCREAMING_SNAKE_CASE_: Optional[int] = torch.exp(0.5 * variance) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler SCREAMING_SNAKE_CASE_: Optional[int] = variance.log() SCREAMING_SNAKE_CASE_: Union[str, Any] = beta.log() SCREAMING_SNAKE_CASE_: Tuple = (predicted_variance + 1) / 2 SCREAMING_SNAKE_CASE_: Optional[Any] = frac * max_log + (1 - frac) * min_log return variance def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : int , lowerCAmelCase__ : Optional[Any] = None , lowerCAmelCase__ : List[str]=None , lowerCAmelCase__ : Tuple = True , ): SCREAMING_SNAKE_CASE_: int = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = torch.split(lowerCAmelCase__ , sample.shape[1] , dim=1) else: SCREAMING_SNAKE_CASE_: List[Any] = None # 1. compute alphas, betas if prev_timestep is None: SCREAMING_SNAKE_CASE_: Optional[Any] = t - 1 SCREAMING_SNAKE_CASE_: Optional[Any] = self.alphas_cumprod[t] SCREAMING_SNAKE_CASE_: Tuple = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one SCREAMING_SNAKE_CASE_: Any = 1 - alpha_prod_t SCREAMING_SNAKE_CASE_: int = 1 - alpha_prod_t_prev if prev_timestep == t - 1: SCREAMING_SNAKE_CASE_: List[str] = self.betas[t] SCREAMING_SNAKE_CASE_: str = self.alphas[t] else: SCREAMING_SNAKE_CASE_: Optional[Any] = 1 - alpha_prod_t / alpha_prod_t_prev SCREAMING_SNAKE_CASE_: Optional[Any] = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": SCREAMING_SNAKE_CASE_: int = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": SCREAMING_SNAKE_CASE_: Optional[Any] = model_output else: raise ValueError( F"prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`" " for the UnCLIPScheduler.") # 3. Clip "predicted x_0" if self.config.clip_sample: SCREAMING_SNAKE_CASE_: Any = torch.clamp( lowerCAmelCase__ , -self.config.clip_sample_range , self.config.clip_sample_range) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf SCREAMING_SNAKE_CASE_: int = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t SCREAMING_SNAKE_CASE_: Union[str, Any] = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf SCREAMING_SNAKE_CASE_: Optional[int] = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise SCREAMING_SNAKE_CASE_: str = 0 if t > 0: SCREAMING_SNAKE_CASE_: int = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=lowerCAmelCase__ , device=model_output.device) SCREAMING_SNAKE_CASE_: List[Any] = self._get_variance( lowerCAmelCase__ , predicted_variance=lowerCAmelCase__ , prev_timestep=lowerCAmelCase__ , ) if self.variance_type == "fixed_small_log": SCREAMING_SNAKE_CASE_: Union[str, Any] = variance elif self.variance_type == "learned_range": SCREAMING_SNAKE_CASE_: int = (0.5 * variance).exp() else: raise ValueError( F"variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`" " for the UnCLIPScheduler.") SCREAMING_SNAKE_CASE_: str = variance * variance_noise SCREAMING_SNAKE_CASE_: Union[str, Any] = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Tuple , ): # Make sure alphas_cumprod and timestep have same device and dtype as original_samples SCREAMING_SNAKE_CASE_: List[str] = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype) SCREAMING_SNAKE_CASE_: Optional[Any] = timesteps.to(original_samples.device) SCREAMING_SNAKE_CASE_: Union[str, Any] = alphas_cumprod[timesteps] ** 0.5 SCREAMING_SNAKE_CASE_: Union[str, Any] = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape) < len(original_samples.shape): SCREAMING_SNAKE_CASE_: str = sqrt_alpha_prod.unsqueeze(-1) SCREAMING_SNAKE_CASE_: List[Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 SCREAMING_SNAKE_CASE_: Optional[int] = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape) < len(original_samples.shape): SCREAMING_SNAKE_CASE_: int = sqrt_one_minus_alpha_prod.unsqueeze(-1) SCREAMING_SNAKE_CASE_: int = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase ) -> list: if len(UpperCAmelCase ) <= 1: return [tuple(UpperCAmelCase )] snake_case_ = [] def generate(UpperCAmelCase , UpperCAmelCase ): snake_case_ = [0] * n res.append(tuple(UpperCAmelCase ) ) snake_case_ = 0 while i < n: if c[i] < i: if i % 2 == 0: snake_case_ , snake_case_ = arr[i], arr[0] else: snake_case_ , snake_case_ = arr[i], arr[c[i]] res.append(tuple(UpperCAmelCase ) ) c[i] += 1 snake_case_ = 0 else: snake_case_ = 0 i += 1 generate(len(UpperCAmelCase ) , UpperCAmelCase ) return res if __name__ == "__main__": __UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
69
0
from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): __lowerCamelCase : List[Any] =DistilBertTokenizer __lowerCamelCase : Union[str, Any] =DistilBertTokenizerFast __lowerCamelCase : Tuple =True @slow def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' __a = DistilBertTokenizer.from_pretrained("""distilbert-base-uncased""" ) __a = tokenizer.encode("""sequence builders""" , add_special_tokens=lowerCAmelCase__ ) __a = tokenizer.encode("""multi-sequence build""" , add_special_tokens=lowerCAmelCase__ ) __a = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ ) __a = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __UpperCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __UpperCamelCase = TaTokenizerFast __UpperCamelCase = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __UpperCamelCase = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
69
0
'''simple docstring''' def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Optional[int] ) -> bool: return str(_UpperCAmelCase ) == str(_UpperCAmelCase )[::-1] def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : List[str] ) -> int: return int(_UpperCAmelCase ) + int(str(_UpperCAmelCase )[::-1] ) def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Any = 10000 ) -> int: _a : List[str] =[] for num in range(1 ,_UpperCAmelCase ): _a : Any =0 _a : Tuple =num while iterations < 50: _a : Dict =sum_reverse(_UpperCAmelCase ) iterations += 1 if is_palindrome(_UpperCAmelCase ): break else: lychrel_nums.append(_UpperCAmelCase ) return len(_UpperCAmelCase ) if __name__ == "__main__": print(F"{solution() = }")
276
"""simple docstring""" import sys from collections import defaultdict class UpperCamelCase : def __init__( self) -> Optional[int]: snake_case_ = [] def a_ ( self, lowerCAmelCase__) -> Any: return self.node_position[vertex] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = pos def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> str: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: snake_case_ = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: snake_case_ = 2 * start + 1 else: snake_case_ = 2 * start + 2 if heap[smallest_child] < heap[start]: snake_case_ , snake_case_ = heap[smallest_child], positions[smallest_child] snake_case_ , snake_case_ = ( heap[start], positions[start], ) snake_case_ , snake_case_ = temp, tempa snake_case_ = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child], self.get_position(positions[start])) self.set_position(positions[start], lowerCAmelCase__) self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> List[str]: snake_case_ = position[index] while index != 0: snake_case_ = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: snake_case_ = heap[parent] snake_case_ = position[parent] self.set_position(position[parent], lowerCAmelCase__) else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, lowerCAmelCase__) break snake_case_ = parent else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, 0) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = len(lowerCAmelCase__) // 2 - 1 for i in range(lowerCAmelCase__, -1, -1): self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, len(lowerCAmelCase__), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = positions[0] snake_case_ = sys.maxsize self.top_to_bottom(lowerCAmelCase__, 0, len(lowerCAmelCase__), lowerCAmelCase__) return temp def UpperCAmelCase ( UpperCAmelCase ) -> Tuple: snake_case_ = Heap() snake_case_ = [0] * len(UpperCAmelCase ) snake_case_ = [-1] * len(UpperCAmelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph snake_case_ = [] # Heap of Distance of vertices from their neighboring vertex snake_case_ = [] for vertex in range(len(UpperCAmelCase ) ): distance_tv.append(sys.maxsize ) positions.append(UpperCAmelCase ) heap.node_position.append(UpperCAmelCase ) snake_case_ = [] snake_case_ = 1 snake_case_ = sys.maxsize for neighbor, distance in adjacency_list[0]: snake_case_ = 0 snake_case_ = distance heap.heapify(UpperCAmelCase , UpperCAmelCase ) for _ in range(1 , len(UpperCAmelCase ) ): snake_case_ = heap.delete_minimum(UpperCAmelCase , UpperCAmelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) snake_case_ = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(UpperCAmelCase )] ): snake_case_ = distance heap.bottom_to_top( UpperCAmelCase , heap.get_position(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase ) snake_case_ = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > __UpperCamelCase = int(input('''Enter number of edges: ''').strip()) __UpperCamelCase = defaultdict(list) for _ in range(edges_number): __UpperCamelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
69
0
'''simple docstring''' from __future__ import annotations from collections.abc import Generator def _A ( ) -> Generator[int, None, None]: _lowercase : List[Any] = {} _lowercase : List[str] = 2 while True: _lowercase : List[str] = factor_map.pop(snake_case , snake_case ) if factor: _lowercase : List[str] = factor + prime while x in factor_map: x += factor _lowercase : Tuple = factor else: _lowercase : Optional[Any] = prime yield prime prime += 1 def _A ( snake_case = 1E10 ) -> int: _lowercase : Tuple = sieve() _lowercase : Optional[Any] = 1 while True: _lowercase : Tuple = next(snake_case ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(snake_case ) n += 2 if __name__ == "__main__": print(solution())
250
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase = { '''configuration_mctct''': ['''MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MCTCTConfig'''], '''feature_extraction_mctct''': ['''MCTCTFeatureExtractor'''], '''processing_mctct''': ['''MCTCTProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MCTCTForCTC''', '''MCTCTModel''', '''MCTCTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
'''simple docstring''' import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all BART models at https://huggingface.co/models?filter=bart __a = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, } __a = { 'facebook/bart-base': 1_024, 'facebook/bart-large': 1_024, 'facebook/bart-large-mnli': 1_024, 'facebook/bart-large-cnn': 1_024, 'facebook/bart-large-xsum': 1_024, 'yjernite/bart_eli5': 1_024, } @lru_cache() def __UpperCAmelCase ( ): _UpperCAmelCase : Optional[Any] = ( list(range(ord("!" ), ord("~" ) + 1 ) ) + list(range(ord("¡" ), ord("¬" ) + 1 ) ) + list(range(ord("®" ), ord("ÿ" ) + 1 ) ) ) _UpperCAmelCase : Optional[int] = bs[:] _UpperCAmelCase : Tuple = 0 for b in range(2**8 ): if b not in bs: bs.append(a_ ) cs.append(2**8 + n ) n += 1 _UpperCAmelCase : Optional[int] = [chr(a_ ) for n in cs] return dict(zip(a_, a_ ) ) def __UpperCAmelCase ( a_: List[str] ): _UpperCAmelCase : Optional[int] = set() _UpperCAmelCase : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _UpperCAmelCase : Dict = char return pairs class A__ ( lowerCAmelCase__ ): """simple docstring""" UpperCamelCase_ : Optional[Any] = VOCAB_FILES_NAMES UpperCamelCase_ : List[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ : Tuple = ['''input_ids''', '''attention_mask'''] def __init__( self : List[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : List[Any]="replace" , lowerCAmelCase__ : Tuple="<s>" , lowerCAmelCase__ : int="</s>" , lowerCAmelCase__ : str="</s>" , lowerCAmelCase__ : List[str]="<s>" , lowerCAmelCase__ : List[Any]="<unk>" , lowerCAmelCase__ : int="<pad>" , lowerCAmelCase__ : Tuple="<mask>" , lowerCAmelCase__ : List[Any]=False , **lowerCAmelCase__ : Optional[int] , ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase : Tuple = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else bos_token _UpperCAmelCase : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else eos_token _UpperCAmelCase : Optional[int] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else sep_token _UpperCAmelCase : Any = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else cls_token _UpperCAmelCase : List[Any] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else unk_token _UpperCAmelCase : Any = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _UpperCAmelCase : Any = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ , **lowerCAmelCase__ , ) with open(lowerCAmelCase__ , encoding="utf-8" ) as vocab_handle: _UpperCAmelCase : int = json.load(lowerCAmelCase__ ) _UpperCAmelCase : Optional[int] = {v: k for k, v in self.encoder.items()} _UpperCAmelCase : Union[str, Any] = errors # how to handle errors in decoding _UpperCAmelCase : Optional[int] = bytes_to_unicode() _UpperCAmelCase : Optional[int] = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ , encoding="utf-8" ) as merges_handle: _UpperCAmelCase : List[Any] = merges_handle.read().split("\n" )[1:-1] _UpperCAmelCase : str = [tuple(merge.split() ) for merge in bpe_merges] _UpperCAmelCase : Optional[Any] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) _UpperCAmelCase : Dict = {} _UpperCAmelCase : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _UpperCAmelCase : Optional[int] = re.compile(R"\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def _lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" return len(self.encoder ) def _lowerCAmelCase ( self : Any ) -> Optional[Any]: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _lowerCAmelCase ( self : Optional[int] , lowerCAmelCase__ : Dict ) -> Union[str, Any]: """simple docstring""" if token in self.cache: return self.cache[token] _UpperCAmelCase : str = tuple(lowerCAmelCase__ ) _UpperCAmelCase : str = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: _UpperCAmelCase : int = min(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break _UpperCAmelCase , _UpperCAmelCase : Dict = bigram _UpperCAmelCase : Any = [] _UpperCAmelCase : Union[str, Any] = 0 while i < len(lowerCAmelCase__ ): try: _UpperCAmelCase : List[str] = word.index(lowerCAmelCase__ , lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _UpperCAmelCase : int = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _UpperCAmelCase : Any = tuple(lowerCAmelCase__ ) _UpperCAmelCase : Optional[int] = new_word if len(lowerCAmelCase__ ) == 1: break else: _UpperCAmelCase : Optional[Any] = get_pairs(lowerCAmelCase__ ) _UpperCAmelCase : Dict = " ".join(lowerCAmelCase__ ) _UpperCAmelCase : Optional[Any] = word return word def _lowerCAmelCase ( self : str , lowerCAmelCase__ : Union[str, Any] ) -> Dict: """simple docstring""" _UpperCAmelCase : Tuple = [] for token in re.findall(self.pat , lowerCAmelCase__ ): _UpperCAmelCase : Optional[Any] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def _lowerCAmelCase ( self : Optional[Any] , lowerCAmelCase__ : str ) -> List[str]: """simple docstring""" return self.encoder.get(lowerCAmelCase__ , self.encoder.get(self.unk_token ) ) def _lowerCAmelCase ( self : Union[str, Any] , lowerCAmelCase__ : str ) -> int: """simple docstring""" return self.decoder.get(lowerCAmelCase__ ) def _lowerCAmelCase ( self : List[str] , lowerCAmelCase__ : str ) -> Optional[int]: """simple docstring""" _UpperCAmelCase : Optional[int] = "".join(lowerCAmelCase__ ) _UpperCAmelCase : int = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def _lowerCAmelCase ( self : List[str] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Optional[Any] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowerCAmelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _UpperCAmelCase : List[Any] = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) _UpperCAmelCase : str = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ ) + "\n" ) _UpperCAmelCase : Tuple = 0 with open(lowerCAmelCase__ , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) _UpperCAmelCase : Tuple = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : int , lowerCAmelCase__ : Any = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _UpperCAmelCase : Dict = [self.cls_token_id] _UpperCAmelCase : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowerCAmelCase ( self : Optional[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Union[str, Any] = None , lowerCAmelCase__ : Optional[Any] = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def _lowerCAmelCase ( self : Optional[int] , lowerCAmelCase__ : str , lowerCAmelCase__ : List[str] = None ) -> List[int]: """simple docstring""" _UpperCAmelCase : int = [self.sep_token_id] _UpperCAmelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowerCAmelCase ( self : int , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Any=False , **lowerCAmelCase__ : Optional[int] ) -> int: """simple docstring""" _UpperCAmelCase : List[Any] = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): _UpperCAmelCase : Any = " " + text return (text, kwargs)
145
"""simple docstring""" from math import factorial def UpperCAmelCase ( UpperCAmelCase = 20 ) -> int: snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(UpperCAmelCase ) / (factorial(UpperCAmelCase ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: __UpperCamelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
69
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """facebook/wav2vec2-base-960h""": """https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json""", # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''wav2vec2''' def __init__( self : Optional[int] , _A : str=32 , _A : List[Any]=768 , _A : Dict=12 , _A : List[Any]=12 , _A : Tuple=3072 , _A : Union[str, Any]="gelu" , _A : str=0.1 , _A : str=0.1 , _A : Optional[int]=0.1 , _A : str=0.0 , _A : Optional[int]=0.0 , _A : Union[str, Any]=0.1 , _A : Optional[int]=0.1 , _A : List[Any]=0.02 , _A : int=1e-5 , _A : int="group" , _A : Tuple="gelu" , _A : Tuple=(512, 512, 512, 512, 512, 512, 512) , _A : List[str]=(5, 2, 2, 2, 2, 2, 2) , _A : List[Any]=(10, 3, 3, 3, 3, 2, 2) , _A : Any=False , _A : List[Any]=128 , _A : Union[str, Any]=16 , _A : Optional[Any]=False , _A : List[Any]=True , _A : int=0.05 , _A : Optional[Any]=10 , _A : Dict=2 , _A : Optional[int]=0.0 , _A : str=10 , _A : List[str]=0 , _A : Optional[Any]=320 , _A : List[Any]=2 , _A : Dict=0.1 , _A : List[str]=100 , _A : Optional[int]=256 , _A : List[str]=256 , _A : Optional[Any]=0.1 , _A : str="sum" , _A : str=False , _A : Dict=False , _A : Optional[int]=256 , _A : int=(512, 512, 512, 512, 1500) , _A : List[str]=(5, 3, 3, 1, 1) , _A : Optional[Any]=(1, 2, 3, 1, 1) , _A : Optional[Any]=512 , _A : int=0 , _A : List[str]=1 , _A : List[str]=2 , _A : str=False , _A : int=3 , _A : Dict=2 , _A : int=3 , _A : Union[str, Any]=None , _A : List[str]=None , **_A : Any , ): """simple docstring""" super().__init__(**lowerCAmelCase__ , pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = hidden_size __SCREAMING_SNAKE_CASE : List[Any] = feat_extract_norm __SCREAMING_SNAKE_CASE : Tuple = feat_extract_activation __SCREAMING_SNAKE_CASE : str = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = conv_bias __SCREAMING_SNAKE_CASE : int = num_conv_pos_embeddings __SCREAMING_SNAKE_CASE : Union[str, Any] = num_conv_pos_embedding_groups __SCREAMING_SNAKE_CASE : str = len(self.conv_dim ) __SCREAMING_SNAKE_CASE : Optional[Any] = num_hidden_layers __SCREAMING_SNAKE_CASE : Tuple = intermediate_size __SCREAMING_SNAKE_CASE : Tuple = hidden_act __SCREAMING_SNAKE_CASE : Union[str, Any] = num_attention_heads __SCREAMING_SNAKE_CASE : Dict = hidden_dropout __SCREAMING_SNAKE_CASE : int = attention_dropout __SCREAMING_SNAKE_CASE : str = activation_dropout __SCREAMING_SNAKE_CASE : Optional[Any] = feat_proj_dropout __SCREAMING_SNAKE_CASE : Optional[Any] = final_dropout __SCREAMING_SNAKE_CASE : Optional[int] = layerdrop __SCREAMING_SNAKE_CASE : Optional[int] = layer_norm_eps __SCREAMING_SNAKE_CASE : Dict = initializer_range __SCREAMING_SNAKE_CASE : List[str] = vocab_size __SCREAMING_SNAKE_CASE : Optional[int] = do_stable_layer_norm __SCREAMING_SNAKE_CASE : Union[str, Any] = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __SCREAMING_SNAKE_CASE : int = apply_spec_augment __SCREAMING_SNAKE_CASE : Optional[int] = mask_time_prob __SCREAMING_SNAKE_CASE : Any = mask_time_length __SCREAMING_SNAKE_CASE : Optional[int] = mask_time_min_masks __SCREAMING_SNAKE_CASE : Optional[Any] = mask_feature_prob __SCREAMING_SNAKE_CASE : List[str] = mask_feature_length __SCREAMING_SNAKE_CASE : Any = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __SCREAMING_SNAKE_CASE : Union[str, Any] = num_codevectors_per_group __SCREAMING_SNAKE_CASE : Dict = num_codevector_groups __SCREAMING_SNAKE_CASE : Optional[int] = contrastive_logits_temperature __SCREAMING_SNAKE_CASE : List[str] = feat_quantizer_dropout __SCREAMING_SNAKE_CASE : Tuple = num_negatives __SCREAMING_SNAKE_CASE : str = codevector_dim __SCREAMING_SNAKE_CASE : Optional[Any] = proj_codevector_dim __SCREAMING_SNAKE_CASE : Optional[Any] = diversity_loss_weight # ctc loss __SCREAMING_SNAKE_CASE : Union[str, Any] = ctc_loss_reduction __SCREAMING_SNAKE_CASE : Dict = ctc_zero_infinity # adapter __SCREAMING_SNAKE_CASE : Any = add_adapter __SCREAMING_SNAKE_CASE : Tuple = adapter_kernel_size __SCREAMING_SNAKE_CASE : List[Any] = adapter_stride __SCREAMING_SNAKE_CASE : Any = num_adapter_layers __SCREAMING_SNAKE_CASE : Tuple = output_hidden_size or hidden_size __SCREAMING_SNAKE_CASE : Union[str, Any] = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. __SCREAMING_SNAKE_CASE : int = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __SCREAMING_SNAKE_CASE : Dict = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Dict = list(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = xvector_output_dim @property def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
303
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
0
"""simple docstring""" def __magic_name__ ( __snake_case : str ) -> int: lowercase : Any = [[0 for _ in range(__snake_case )] for _ in range(m + 1 )] for i in range(m + 1 ): lowercase : str = 1 for n in range(m + 1 ): for k in range(1 , __snake_case ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: _A : Tuple = int(input("""Enter a number: """).strip()) print(partition(n)) except ValueError: print("""Please enter a number.""") else: try: _A : Tuple = int(sys.argv[1]) print(partition(n)) except ValueError: print("""Please pass a number.""")
202
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = (DPMSolverSinglestepScheduler,) SCREAMING_SNAKE_CASE_ = (("num_inference_steps", 2_5),) def a_ ( self, **lowerCAmelCase__) -> int: snake_case_ = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf'), 'variance_type': None, } config.update(**lowerCAmelCase__) return config def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> List[Any]: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ , snake_case_ = sample, sample for t in range(lowerCAmelCase__, time_step + scheduler.config.solver_order + 1): snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self) -> Union[str, Any]: pass def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> int: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self, lowerCAmelCase__=None, **lowerCAmelCase__) -> Union[str, Any]: if scheduler is None: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample return sample def a_ ( self) -> List[Any]: snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = 50 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2574) < 1e-3 def a_ ( self) -> Dict: for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__) def a_ ( self) -> Optional[Any]: # make sure that iterating over schedulers with same config names gives same results # for defaults snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 snake_case_ = DEISMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverMultistepScheduler.from_config(scheduler.config) snake_case_ = UniPCMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverSinglestepScheduler.from_config(scheduler.config) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> str: self.check_over_configs(thresholding=lowerCAmelCase__) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCAmelCase__, prediction_type=lowerCAmelCase__, sample_max_value=lowerCAmelCase__, algorithm_type='dpmsolver++', solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, ) def a_ ( self) -> Tuple: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__) def a_ ( self) -> Optional[int]: for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) snake_case_ = self.full_loop( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) assert not torch.isnan(lowerCAmelCase__).any(), "Samples have nan numbers" def a_ ( self) -> Optional[Any]: self.check_over_configs(lower_order_final=lowerCAmelCase__) self.check_over_configs(lower_order_final=lowerCAmelCase__) def a_ ( self) -> Any: self.check_over_configs(lambda_min_clipped=-float('inf')) self.check_over_configs(lambda_min_clipped=-5.1) def a_ ( self) -> Any: self.check_over_configs(variance_type=lowerCAmelCase__) self.check_over_configs(variance_type='learned_range') def a_ ( self) -> List[Any]: for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=lowerCAmelCase__, time_step=0) def a_ ( self) -> int: snake_case_ = self.full_loop() snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> Dict: snake_case_ = self.full_loop(use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2248) < 1e-3 def a_ ( self) -> Union[str, Any]: snake_case_ = self.full_loop(prediction_type='v_prediction') snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.1453) < 1e-3 def a_ ( self) -> Optional[Any]: snake_case_ = self.full_loop(prediction_type='v_prediction', use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.0649) < 1e-3 def a_ ( self) -> Optional[int]: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(thresholding=lowerCAmelCase__, dynamic_thresholding_ratio=0) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample assert sample.dtype == torch.floataa
69
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging lowercase : Any = logging.get_logger(__name__) lowercase : int = """▁""" lowercase : str = {"""vocab_file""": """sentencepiece.bpe.model"""} lowercase : Dict = { """vocab_file""": { """facebook/mbart-large-50-one-to-many-mmt""": ( """https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model""" ), } } lowercase : Union[str, Any] = { """facebook/mbart-large-50-one-to-many-mmt""": 1_0_2_4, } # fmt: off lowercase : Union[str, Any] = ["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN""", """af_ZA""", """az_AZ""", """bn_IN""", """fa_IR""", """he_IL""", """hr_HR""", """id_ID""", """ka_GE""", """km_KH""", """mk_MK""", """ml_IN""", """mn_MN""", """mr_IN""", """pl_PL""", """ps_AF""", """pt_XX""", """sv_SE""", """sw_KE""", """ta_IN""", """te_IN""", """th_TH""", """tl_XX""", """uk_UA""", """ur_PK""", """xh_ZA""", """gl_ES""", """sl_SI"""] class A__ ( lowerCAmelCase__ ): """simple docstring""" __A : Dict = VOCAB_FILES_NAMES __A : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : List[Any] = PRETRAINED_VOCAB_FILES_MAP __A : int = ['''input_ids''', '''attention_mask'''] __A : Optional[int] = [] __A : str = [] def __init__( self , lowercase , lowercase=None , lowercase=None , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase = None , **lowercase , ) -> None: '''simple docstring''' a__ : Union[str, Any] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__ , lowerCAmelCase__) else mask_token a__ : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs a__ : Union[str, Any] = kwargs.get('additional_special_tokens' , []) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase__ , ) a__ : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowerCAmelCase__)) a__ : List[str] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token a__ : Optional[Any] = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab a__ : Tuple = 1 a__ : int = len(self.sp_model) a__ : Optional[int] = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase__) } a__ : Optional[Any] = {v: k for k, v in self.lang_code_to_id.items()} a__ : Any = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) a__ : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} a__ : Optional[Any] = src_lang if src_lang is not None else 'en_XX' a__ : Optional[Any] = self.lang_code_to_id[self._src_lang] a__ : Union[str, Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def __lowercase ( self) -> int: '''simple docstring''' return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __lowercase ( self) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def __lowercase ( self , lowercase) -> None: '''simple docstring''' a__ : Dict = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self) -> Dict: '''simple docstring''' a__ : List[str] = self.__dict__.copy() a__ : List[str] = None return state def __setstate__( self , lowercase) -> None: '''simple docstring''' a__ : Tuple = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs'): a__ : Optional[int] = {} a__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __lowercase ( self) -> Dict: '''simple docstring''' a__ : int = {self.convert_ids_to_tokens(lowerCAmelCase__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __lowercase ( self , lowercase) -> List[str]: '''simple docstring''' return self.sp_model.encode(lowerCAmelCase__ , out_type=lowerCAmelCase__) def __lowercase ( self , lowercase) -> int: '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] a__ : Dict = self.sp_model.PieceToId(lowerCAmelCase__) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __lowercase ( self , lowercase) -> str: '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def __lowercase ( self , lowercase) -> Tuple: '''simple docstring''' a__ : List[str] = [] a__ : int = '' a__ : Optional[Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCAmelCase__) + token a__ : List[str] = True a__ : Any = [] else: current_sub_tokens.append(lowerCAmelCase__) a__ : Optional[Any] = False out_string += self.sp_model.decode(lowerCAmelCase__) return out_string.strip() def __lowercase ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return a__ : Tuple = os.path.join( lowerCAmelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowerCAmelCase__) elif not os.path.isfile(self.vocab_file): with open(lowerCAmelCase__ , 'wb') as fi: a__ : int = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__) return (out_vocab_file,) def __lowercase ( self , lowercase , lowercase = None , lowercase = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__) a__ : Optional[int] = [1] * len(self.prefix_tokens) a__ : int = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__)) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__)) + ([0] * len(lowerCAmelCase__)) + suffix_ones def __lowercase ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowercase ( self , lowercase , lowercase , lowercase , lowercase , **lowercase) -> Optional[int]: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') a__ : Optional[int] = src_lang a__ : Optional[Any] = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__) a__ : List[Any] = self.convert_tokens_to_ids(lowerCAmelCase__) a__ : Optional[int] = tgt_lang_id return inputs def __lowercase ( self , lowercase , lowercase = "en_XX" , lowercase = None , lowercase = "ro_RO" , **lowercase , ) -> BatchEncoding: '''simple docstring''' a__ : Any = src_lang a__ : Union[str, Any] = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__) def __lowercase ( self) -> Optional[int]: '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang) def __lowercase ( self) -> Tuple: '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang) def __lowercase ( self , lowercase) -> None: '''simple docstring''' a__ : Optional[Any] = self.lang_code_to_id[src_lang] a__ : Union[str, Any] = [self.cur_lang_code_id] a__ : List[str] = [self.eos_token_id] def __lowercase ( self , lowercase) -> None: '''simple docstring''' a__ : Tuple = self.lang_code_to_id[tgt_lang] a__ : Any = [self.cur_lang_code_id] a__ : List[Any] = [self.eos_token_id]
99
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # Base Case if curr_ind == len(UpperCAmelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(UpperCAmelCase ) ): if valid_connection(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # Insert current vertex into path as next transition snake_case_ = next_ver # Validate created path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , curr_ind + 1 ): return True # Backtrack snake_case_ = -1 return False def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = 0 ) -> list[int]: snake_case_ = [-1] * (len(UpperCAmelCase ) + 1) # initialize start and end of path with starting index snake_case_ = snake_case_ = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , 1 ) else []
69
0
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py A_ : List[Any] = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. A_ : Dict = direct_transformers_import(PATH_TO_TRANSFORMERS) A_ : int = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` A_ : str = re.compile(R'\[(.+?)\]\((https://huggingface\.co/.+?)\)') A_ : Union[str, Any] = { 'DecisionTransformerConfig', 'EncoderDecoderConfig', 'MusicgenConfig', 'RagConfig', 'SpeechEncoderDecoderConfig', 'TimmBackboneConfig', 'VisionEncoderDecoderConfig', 'VisionTextDualEncoderConfig', 'LlamaConfig', } def __a ( SCREAMING_SNAKE_CASE ) -> List[Any]: '''simple docstring''' __UpperCAmelCase = None # source code of `config_class` __UpperCAmelCase = inspect.getsource(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = _re_checkpoint.findall(SCREAMING_SNAKE_CASE ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): __UpperCAmelCase = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link __UpperCAmelCase = f'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: __UpperCAmelCase = ckpt_name break return checkpoint def __a ( ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue __UpperCAmelCase = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: __UpperCAmelCase = '''\n'''.join(sorted(SCREAMING_SNAKE_CASE ) ) raise ValueError(f'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
333
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''▁''' __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model''' ), } } __UpperCamelCase = { '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__ = None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = {} if sp_model_kwargs is None else sp_model_kwargs snake_case_ = legacy_behaviour super().__init__( bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, sp_model_kwargs=self.sp_model_kwargs, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowerCAmelCase__)) snake_case_ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token snake_case_ = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab snake_case_ = 1 snake_case_ = len(self.sp_model) snake_case_ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase__) } snake_case_ = {v: k for k, v in self.lang_code_to_id.items()} snake_case_ = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) snake_case_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} snake_case_ = list(self.lang_code_to_id.keys()) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens]) snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.lang_code_to_id[self._src_lang] snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self) -> Union[str, Any]: snake_case_ = self.__dict__.copy() snake_case_ = None snake_case_ = self.sp_model.serialized_model_proto() return state def __setstate__( self, lowerCAmelCase__) -> Tuple: snake_case_ = d # for backward compatibility if not hasattr(self, 'sp_model_kwargs'): snake_case_ = {} snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) @property def a_ ( self) -> str: return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None, lowerCAmelCase__ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__, token_ids_a=lowerCAmelCase__, already_has_special_tokens=lowerCAmelCase__) snake_case_ = [1] * len(self.prefix_tokens) snake_case_ = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__)) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__)) + ([0] * len(lowerCAmelCase__)) + suffix_ones def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self) -> List[Any]: snake_case_ = {self.convert_ids_to_tokens(lowerCAmelCase__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def a_ ( self, lowerCAmelCase__) -> List[str]: return self.sp_model.encode(lowerCAmelCase__, out_type=lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> Any: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case_ = self.sp_model.PieceToId(lowerCAmelCase__) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a_ ( self, lowerCAmelCase__) -> Dict: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def a_ ( self, lowerCAmelCase__) -> List[str]: snake_case_ = ''.join(lowerCAmelCase__).replace(lowerCAmelCase__, ' ').strip() return out_string def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file, lowerCAmelCase__) elif not os.path.isfile(self.vocab_file): with open(lowerCAmelCase__, 'wb') as fi: snake_case_ = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__) return (out_vocab_file,) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> Union[str, Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> int: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[src_lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id]
69
0
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (lowerCAmelCase__ ): """simple docstring""" def __init__( self : str , *lowerCamelCase : Optional[Any] , **lowerCamelCase : Optional[int] ) -> None: warnings.warn( "The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use MobileViTImageProcessor instead." , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
123
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCAmelCase ( ) -> int: snake_case_ = HfArgumentParser(UpperCAmelCase ) snake_case_ = parser.parse_args_into_dataclasses()[0] snake_case_ = TensorFlowBenchmark(args=UpperCAmelCase ) try: snake_case_ = parser.parse_args_into_dataclasses()[0] except ValueError as e: snake_case_ = 'Arg --no_{0} is no longer used, please use --no-{0} instead.' snake_case_ = ' '.join(str(UpperCAmelCase ).split(' ' )[:-1] ) snake_case_ = '' snake_case_ = eval(str(UpperCAmelCase ).split(' ' )[-1] ) snake_case_ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = full_error_msg + begin_error_msg + str(UpperCAmelCase ) raise ValueError(UpperCAmelCase ) benchmark.run() if __name__ == "__main__": main()
69
0
import qiskit def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> qiskit.result.counts.Counts: lowerCamelCase : List[str] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register lowerCamelCase : List[Any] = qiskit.QuantumCircuit(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # Apply X (NOT) Gate to Qubits 0 & 1 circuit.x(0 ) circuit.x(1 ) # Map the quantum measurement to the classical bits circuit.measure([0, 1] ,[0, 1] ) # Execute the circuit on the qasm simulator lowerCamelCase : List[str] = qiskit.execute(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Dict = single_qubit_measure(2, 2) print(f'''Total count for various states are: {counts}''')
48
"""simple docstring""" from __future__ import annotations def UpperCAmelCase ( UpperCAmelCase ) -> None: create_state_space_tree(UpperCAmelCase , [] , 0 , [0 for i in range(len(UpperCAmelCase ) )] ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) -> None: if index == len(UpperCAmelCase ): print(UpperCAmelCase ) return for i in range(len(UpperCAmelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) snake_case_ = True create_state_space_tree(UpperCAmelCase , UpperCAmelCase , index + 1 , UpperCAmelCase ) current_sequence.pop() snake_case_ = False __UpperCamelCase = [3, 1, 2, 4] generate_all_permutations(sequence) __UpperCamelCase = ["A", "B", "C"] generate_all_permutations(sequence_a)
69
0
from __future__ import annotations from collections.abc import Iterator from typing import Any class __lowercase : """simple docstring""" def __init__( self : Optional[int] , lowerCAmelCase__ : Tuple): SCREAMING_SNAKE_CASE_: Dict = data SCREAMING_SNAKE_CASE_: Union[str, Any] = None class __lowercase : """simple docstring""" def __init__( self : Optional[int]): SCREAMING_SNAKE_CASE_: List[str] = None SCREAMING_SNAKE_CASE_: Optional[Any] = None def __iter__( self : Any): SCREAMING_SNAKE_CASE_: Tuple = self.head while self.head: yield node.data SCREAMING_SNAKE_CASE_: Dict = node.next if node == self.head: break def __len__( self : Tuple): return sum(1 for _ in self) def __repr__( self : Optional[Any]): return "->".join(str(lowerCAmelCase__) for item in iter(self)) def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase__ : str): self.insert_nth(len(self) , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase__ : Tuple): self.insert_nth(0 , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Any): if index < 0 or index > len(self): raise IndexError("list index out of range.") SCREAMING_SNAKE_CASE_: Any = Node(lowerCAmelCase__) if self.head is None: SCREAMING_SNAKE_CASE_: Optional[Any] = new_node # first node points itself SCREAMING_SNAKE_CASE_: List[Any] = new_node elif index == 0: # insert at head SCREAMING_SNAKE_CASE_: List[Any] = self.head SCREAMING_SNAKE_CASE_: Dict = new_node else: SCREAMING_SNAKE_CASE_: List[str] = self.head for _ in range(index - 1): SCREAMING_SNAKE_CASE_: List[Any] = temp.next SCREAMING_SNAKE_CASE_: Dict = temp.next SCREAMING_SNAKE_CASE_: str = new_node if index == len(self) - 1: # insert at tail SCREAMING_SNAKE_CASE_: Union[str, Any] = new_node def _SCREAMING_SNAKE_CASE ( self : str): return self.delete_nth(0) def _SCREAMING_SNAKE_CASE ( self : str): return self.delete_nth(len(self) - 1) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase__ : Union[str, Any] = 0): if not 0 <= index < len(self): raise IndexError("list index out of range.") SCREAMING_SNAKE_CASE_: Union[str, Any] = self.head if self.head == self.tail: # just one node SCREAMING_SNAKE_CASE_: Dict = None elif index == 0: # delete head node SCREAMING_SNAKE_CASE_: List[Any] = self.tail.next.next SCREAMING_SNAKE_CASE_: List[Any] = self.head.next else: SCREAMING_SNAKE_CASE_: List[str] = self.head for _ in range(index - 1): SCREAMING_SNAKE_CASE_: Optional[Any] = temp.next SCREAMING_SNAKE_CASE_: Union[str, Any] = temp.next SCREAMING_SNAKE_CASE_: Optional[int] = temp.next.next if index == len(self) - 1: # delete at tail SCREAMING_SNAKE_CASE_: str = temp return delete_node.data def _SCREAMING_SNAKE_CASE ( self : Tuple): return len(self) == 0 def A_ ( ): SCREAMING_SNAKE_CASE_: Dict = CircularLinkedList() assert len(_UpperCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(_UpperCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(_UpperCAmelCase ) == i circular_linked_list.insert_nth(_UpperCAmelCase , i + 1 ) assert str(_UpperCAmelCase ) == "->".join(str(_UpperCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(_UpperCAmelCase ) == "->".join(str(_UpperCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(_UpperCAmelCase ) == "->".join(str(_UpperCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(_UpperCAmelCase ) == "->".join(str(_UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(_UpperCAmelCase ) == "->".join(str(_UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( 'Warning: upper bound of deterministic test is exceeded. ' 'Pass allow_probable=True to allow probabilistic test. ' 'A return value of True indicates a probable prime.' ) # array bounds provided by analysis snake_case_ = [ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] snake_case_ = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(UpperCAmelCase , 1 ): if n < _p: # then we have our last prime to check snake_case_ = primes[:idx] break snake_case_ , snake_case_ = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: snake_case_ = False for r in range(UpperCAmelCase ): snake_case_ = pow(UpperCAmelCase , d * 2**r , UpperCAmelCase ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): snake_case_ = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def UpperCAmelCase ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
69
0
import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class SCREAMING_SNAKE_CASE ( unittest.TestCase ): def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase_ ( self : str ): '''simple docstring''' __a = StableDiffusionKDiffusionPipeline.from_pretrained("""CompVis/stable-diffusion-v1-4""" ) __a = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) sd_pipe.set_scheduler("""sample_euler""" ) __a = """A painting of a squirrel eating a burger""" __a = torch.manual_seed(0 ) __a = sd_pipe([prompt] , generator=lowerCAmelCase__ , guidance_scale=9.0 , num_inference_steps=20 , output_type="""np""" ) __a = output.images __a = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __a = np.array([0.0447, 0.0492, 0.0468, 0.0408, 0.0383, 0.0408, 0.0354, 0.0380, 0.0339] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : int ): '''simple docstring''' __a = StableDiffusionKDiffusionPipeline.from_pretrained("""stabilityai/stable-diffusion-2-1-base""" ) __a = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) sd_pipe.set_scheduler("""sample_euler""" ) __a = """A painting of a squirrel eating a burger""" __a = torch.manual_seed(0 ) __a = sd_pipe([prompt] , generator=lowerCAmelCase__ , guidance_scale=9.0 , num_inference_steps=20 , output_type="""np""" ) __a = output.images __a = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __a = np.array([0.1237, 0.1320, 0.1438, 0.1359, 0.1390, 0.1132, 0.1277, 0.1175, 0.1112] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' __a = StableDiffusionKDiffusionPipeline.from_pretrained("""stabilityai/stable-diffusion-2-1-base""" ) __a = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) sd_pipe.set_scheduler("""sample_dpmpp_2m""" ) __a = """A painting of a squirrel eating a burger""" __a = torch.manual_seed(0 ) __a = sd_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=7.5 , num_inference_steps=15 , output_type="""np""" , use_karras_sigmas=lowerCAmelCase__ , ) __a = output.images __a = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __a = np.array( [0.11381689, 0.12112921, 0.1389457, 0.12549606, 0.1244964, 0.10831517, 0.11562866, 0.10867816, 0.10499048] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __UpperCamelCase = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
69
0
'''simple docstring''' from __future__ import annotations import math import numpy as np from numpy.linalg import norm def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Dict ,_UpperCAmelCase : Optional[Any] ) -> float: return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(_UpperCAmelCase ,_UpperCAmelCase ) ) ) def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Optional[int] ,_UpperCAmelCase : Tuple ) -> list[list[list[float] | float]]: if dataset.ndim != value_array.ndim: _a : Dict =( """Wrong input data\'s dimensions... """ F"dataset : {dataset.ndim}, value_array : {value_array.ndim}" ) raise ValueError(_UpperCAmelCase ) try: if dataset.shape[1] != value_array.shape[1]: _a : List[str] =( """Wrong input data\'s shape... """ F"dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}" ) raise ValueError(_UpperCAmelCase ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("""Wrong shape""" ) if dataset.dtype != value_array.dtype: _a : int =( """Input data have different datatype... """ F"dataset : {dataset.dtype}, value_array : {value_array.dtype}" ) raise TypeError(_UpperCAmelCase ) _a : Tuple =[] for value in value_array: _a : Tuple =euclidean(_UpperCAmelCase ,dataset[0] ) _a : str =dataset[0].tolist() for dataset_value in dataset[1:]: _a : List[str] =euclidean(_UpperCAmelCase ,_UpperCAmelCase ) if dist > temp_dist: _a : List[Any] =temp_dist _a : int =dataset_value.tolist() answer.append([vector, dist] ) return answer def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Tuple ,_UpperCAmelCase : Optional[Any] ) -> float: return np.dot(_UpperCAmelCase ,_UpperCAmelCase ) / (norm(_UpperCAmelCase ) * norm(_UpperCAmelCase )) if __name__ == "__main__": import doctest doctest.testmod()
276
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: __UpperCamelCase = None __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } __UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = NllbTokenizer SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True snake_case_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) snake_case_ = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__) for lang_code in FAIRSEQ_LANGUAGE_CODES } snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.convert_tokens_to_ids(self._src_lang) snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> Tuple: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__): copyfile(self.vocab_file, lowerCAmelCase__) return (out_vocab_file,)
69
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { 'configuration_nllb_moe': [ 'NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'NllbMoeConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST', 'NllbMoeForConditionalGeneration', 'NllbMoeModel', 'NllbMoePreTrainedModel', 'NllbMoeTop2Router', 'NllbMoeSparseMLP', ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
250
"""simple docstring""" from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''T5Config''' class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig
69
0
'''simple docstring''' import math def __UpperCAmelCase ( a_: List[str], a_: Any ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(a_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. __a = 'Enter the base and the power separated by a comma: ' __a , __a = map(int, input(prompt).split(',')) __a , __a = map(int, input(prompt).split(',')) # We find the log of each number, using the function res(), which takes two # arguments. __a = res(xa, ya) __a = res(xa, ya) # We check for the largest number if resa > resa: print('Largest number is', xa, '^', ya) elif resa > resa: print('Largest number is', xa, '^', ya) else: print('Both are equal')
145
"""simple docstring""" import argparse __UpperCamelCase = '''docs/source/_static/js/custom.js''' def UpperCAmelCase ( UpperCAmelCase ) -> int: with open(UpperCAmelCase , encoding='utf-8' , newline='\n' ) as f: snake_case_ = f.readlines() snake_case_ = 0 # First let's put the right version while not lines[index].startswith('const stableVersion =' ): index += 1 snake_case_ = f'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith('const versionMapping = {' ): index += 1 # We go until the end while not lines[index].startswith('}' ): index += 1 # We add the new version at the end lines[index - 1] += f' "v{version}": "v{version}",\n' with open(UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') __UpperCamelCase = parser.parse_args() update_custom_js(args.version)
69
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/config.json""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/config.json""", """funnel-transformer/medium-base""": """https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json""", """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/config.json""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json""", """funnel-transformer/xlarge-base""": """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json""", } class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''funnel''' lowerCAmelCase_ = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''n_head''', } def __init__( self : Tuple , _A : Tuple=3_0522 , _A : Union[str, Any]=[4, 4, 4] , _A : List[Any]=None , _A : Union[str, Any]=2 , _A : Dict=768 , _A : Tuple=12 , _A : Tuple=64 , _A : int=3072 , _A : int="gelu_new" , _A : int=0.1 , _A : List[Any]=0.1 , _A : str=0.0 , _A : Dict=0.1 , _A : Any=None , _A : List[str]=1e-9 , _A : int="mean" , _A : Optional[int]="relative_shift" , _A : Optional[int]=True , _A : List[str]=True , _A : str=True , **_A : Any , ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = vocab_size __SCREAMING_SNAKE_CASE : int = block_sizes __SCREAMING_SNAKE_CASE : List[str] = [1] * len(lowerCAmelCase__ ) if block_repeats is None else block_repeats assert len(lowerCAmelCase__ ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." __SCREAMING_SNAKE_CASE : int = num_decoder_layers __SCREAMING_SNAKE_CASE : Any = d_model __SCREAMING_SNAKE_CASE : Dict = n_head __SCREAMING_SNAKE_CASE : Optional[int] = d_head __SCREAMING_SNAKE_CASE : Any = d_inner __SCREAMING_SNAKE_CASE : Optional[int] = hidden_act __SCREAMING_SNAKE_CASE : List[Any] = hidden_dropout __SCREAMING_SNAKE_CASE : Tuple = attention_dropout __SCREAMING_SNAKE_CASE : List[Any] = activation_dropout __SCREAMING_SNAKE_CASE : Tuple = initializer_range __SCREAMING_SNAKE_CASE : Tuple = initializer_std __SCREAMING_SNAKE_CASE : Any = layer_norm_eps assert pooling_type in [ "mean", "max", ], F'''Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.''' __SCREAMING_SNAKE_CASE : List[Any] = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F'''Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.''' __SCREAMING_SNAKE_CASE : str = attention_type __SCREAMING_SNAKE_CASE : Any = separate_cls __SCREAMING_SNAKE_CASE : int = truncate_seq __SCREAMING_SNAKE_CASE : List[Any] = pool_q_only super().__init__(**lowerCAmelCase__ ) @property def UpperCAmelCase__ ( self : Dict ): """simple docstring""" return sum(self.block_sizes ) @num_hidden_layers.setter def UpperCAmelCase__ ( self : Union[str, Any] , _A : Union[str, Any] ): """simple docstring""" raise NotImplementedError( '''This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.''' ) @property def UpperCAmelCase__ ( self : Dict ): """simple docstring""" return len(self.block_sizes ) @num_blocks.setter def UpperCAmelCase__ ( self : Optional[int] , _A : int ): """simple docstring""" raise NotImplementedError('''This model does not support the setting of `num_blocks`. Please set `block_sizes`.''' )
303
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class UpperCamelCase : def __init__( self, lowerCAmelCase__) -> Optional[int]: snake_case_ = data snake_case_ = None class UpperCamelCase : def __init__( self) -> Dict: snake_case_ = None snake_case_ = None def __iter__( self) -> Iterator[Any]: snake_case_ = self.head while self.head: yield node.data snake_case_ = node.next if node == self.head: break def __len__( self) -> int: return sum(1 for _ in self) def __repr__( self) -> str: return "->".join(str(lowerCAmelCase__) for item in iter(self)) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(len(self), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(0, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> None: if index < 0 or index > len(self): raise IndexError('list index out of range.') snake_case_ = Node(lowerCAmelCase__) if self.head is None: snake_case_ = new_node # first node points itself snake_case_ = snake_case_ = new_node elif index == 0: # insert at head snake_case_ = self.head snake_case_ = snake_case_ = new_node else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = new_node if index == len(self) - 1: # insert at tail snake_case_ = new_node def a_ ( self) -> str: return self.delete_nth(0) def a_ ( self) -> Any: return self.delete_nth(len(self) - 1) def a_ ( self, lowerCAmelCase__ = 0) -> Any: if not 0 <= index < len(self): raise IndexError('list index out of range.') snake_case_ = self.head if self.head == self.tail: # just one node snake_case_ = snake_case_ = None elif index == 0: # delete head node snake_case_ = self.tail.next.next snake_case_ = self.head.next else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = temp.next.next if index == len(self) - 1: # delete at tail snake_case_ = temp return delete_node.data def a_ ( self) -> bool: return len(self) == 0 def UpperCAmelCase ( ) -> None: snake_case_ = CircularLinkedList() assert len(UpperCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(UpperCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(UpperCAmelCase ) == i circular_linked_list.insert_nth(UpperCAmelCase , i + 1 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
69
0
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging _A : str = logging.get_logger(__name__) # pylint: disable=invalid-name class a__ ( lowerCAmelCase__ ): def __init__( self , _a , _a=768 ): super().__init__(lowerCAmelCase__ ) lowercase : Tuple = proj_size lowercase : str = CLIPVisionModel(lowerCAmelCase__ ) lowercase : Union[str, Any] = PaintByExampleMapper(lowerCAmelCase__ ) lowercase : int = nn.LayerNorm(config.hidden_size ) lowercase : Dict = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling lowercase : List[Any] = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def __magic_name__ ( self , _a , _a=False ): lowercase : Optional[Any] = self.model(pixel_values=lowerCAmelCase__ ) lowercase : List[Any] = clip_output.pooler_output lowercase : Dict = self.mapper(latent_states[:, None] ) lowercase : int = self.final_layer_norm(lowerCAmelCase__ ) lowercase : str = self.proj_out(lowerCAmelCase__ ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class a__ ( nn.Module ): def __init__( self , _a ): super().__init__() lowercase : str = (config.num_hidden_layers + 1) // 5 lowercase : Union[str, Any] = config.hidden_size lowercase : Dict = 1 lowercase : int = nn.ModuleList( [ BasicTransformerBlock(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , activation_fn="gelu" , attention_bias=lowerCAmelCase__ ) for _ in range(lowerCAmelCase__ ) ] ) def __magic_name__ ( self , _a ): for block in self.blocks: lowercase : List[Any] = block(lowerCAmelCase__ ) return hidden_states
202
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCamelCase = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __UpperCamelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) __UpperCamelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __UpperCamelCase = re.compile(r'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __UpperCamelCase = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def UpperCAmelCase ( UpperCAmelCase ) -> List[Any]: snake_case_ = None # source code of `config_class` snake_case_ = inspect.getsource(UpperCAmelCase ) snake_case_ = _re_checkpoint.findall(UpperCAmelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): snake_case_ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: snake_case_ = ckpt_name break return checkpoint def UpperCAmelCase ( ) -> Union[str, Any]: snake_case_ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ = get_checkpoint_from_config_class(UpperCAmelCase ) snake_case_ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = '\n'.join(sorted(UpperCAmelCase ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
import socket def A_ ( ) -> List[str]: a__ : Any = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) a__ : Dict = socket.gethostname() a__ : List[str] = 1_2312 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: a__ : Any = sock.recv(1024 ) if not data: break out_file.write(A__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
99
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __UpperCamelCase = 0 __UpperCamelCase = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __UpperCamelCase = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __UpperCamelCase = tuple[int, int] class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, ) -> None: snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = g_cost snake_case_ = parent snake_case_ = self.calculate_heuristic() snake_case_ = self.g_cost + self.h_cost def a_ ( self) -> float: snake_case_ = self.pos_x - self.goal_x snake_case_ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase__) + abs(lowerCAmelCase__) else: return sqrt(dy**2 + dx**2) def __lt__( self, lowerCAmelCase__) -> bool: return self.f_cost < other.f_cost class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = Node(start[1], start[0], goal[1], goal[0], 0, lowerCAmelCase__) snake_case_ = Node(goal[1], goal[0], goal[1], goal[0], 9_9999, lowerCAmelCase__) snake_case_ = [self.start] snake_case_ = [] snake_case_ = False def a_ ( self) -> list[TPosition]: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() snake_case_ = self.open_nodes.pop(0) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase__) self.closed_nodes.append(lowerCAmelCase__) snake_case_ = self.get_successors(lowerCAmelCase__) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__) else: self.open_nodes.append(lowerCAmelCase__) return [self.start.pos] def a_ ( self, lowerCAmelCase__) -> list[Node]: snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(lowerCAmelCase__) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__, lowerCAmelCase__, self.target.pos_y, self.target.pos_x, parent.g_cost + 1, lowerCAmelCase__, )) return successors def a_ ( self, lowerCAmelCase__) -> list[TPosition]: snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x)) snake_case_ = current_node.parent path.reverse() return path class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> None: snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = False def a_ ( self) -> list[TPosition]: while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() snake_case_ = self.fwd_astar.open_nodes.pop(0) snake_case_ = self.bwd_astar.open_nodes.pop(0) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase__, lowerCAmelCase__) self.fwd_astar.closed_nodes.append(lowerCAmelCase__) self.bwd_astar.closed_nodes.append(lowerCAmelCase__) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase__), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase__), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase__) else: astar.open_nodes.append(lowerCAmelCase__) return [self.fwd_astar.start.pos] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> list[TPosition]: snake_case_ = self.fwd_astar.retrace_path(lowerCAmelCase__) snake_case_ = self.bwd_astar.retrace_path(lowerCAmelCase__) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __UpperCamelCase = (0, 0) __UpperCamelCase = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __UpperCamelCase = time.time() __UpperCamelCase = AStar(init, goal) __UpperCamelCase = a_star.search() __UpperCamelCase = time.time() - start_time print(F"""AStar execution time = {end_time:f} seconds""") __UpperCamelCase = time.time() __UpperCamelCase = BidirectionalAStar(init, goal) __UpperCamelCase = time.time() - bd_start_time print(F"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
69
0
from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class A_ ( lowerCAmelCase__ ): '''simple docstring''' a__ = 42 a__ = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
333
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: while a != 0: snake_case_ , snake_case_ = b % a, a return b def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: if gcd(UpperCAmelCase , UpperCAmelCase ) != 1: snake_case_ = f'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = 1, 0, a snake_case_ , snake_case_ , snake_case_ = 0, 1, m while va != 0: snake_case_ = ua // va snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
69
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase=False ): __snake_case : List[str] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'module.blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'module.blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append( (F'module.blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'module.blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'module.blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'module.blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'module.blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'module.blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'module.blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'module.blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("module.cls_token", "vit.embeddings.cls_token"), ("module.patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("module.patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("module.pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("module.norm.weight", "layernorm.weight"), ("module.norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __snake_case : int = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: __snake_case : List[str] = "" else: __snake_case : List[Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __snake_case : Any = state_dict.pop(F'module.blocks.{i}.attn.qkv.weight' ) __snake_case : List[str] = state_dict.pop(F'module.blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __snake_case : Dict = in_proj_weight[ : config.hidden_size, : ] __snake_case : Tuple = in_proj_bias[: config.hidden_size] __snake_case : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __snake_case : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __snake_case : List[str] = in_proj_weight[ -config.hidden_size :, : ] __snake_case : Optional[Any] = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): # projection head is used in the self-supervised pre-training in MSN, # for downstream task it's not needed. __snake_case : Tuple = [ "module.fc.fc1.weight", "module.fc.fc1.bias", "module.fc.bn1.weight", "module.fc.bn1.bias", "module.fc.bn1.running_mean", "module.fc.bn1.running_var", "module.fc.bn1.num_batches_tracked", "module.fc.fc2.weight", "module.fc.fc2.bias", "module.fc.bn2.weight", "module.fc.bn2.bias", "module.fc.bn2.running_mean", "module.fc.bn2.running_var", "module.fc.bn2.num_batches_tracked", "module.fc.fc3.weight", "module.fc.fc3.bias", ] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : str = dct.pop(__lowerCamelCase ) __snake_case : Union[str, Any] = val def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = ViTMSNConfig() __snake_case : Optional[int] = 1_0_0_0 __snake_case : Optional[int] = "datasets/huggingface/label-files" __snake_case : Union[str, Any] = "imagenet-1k-id2label.json" __snake_case : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase ) , "r" ) ) __snake_case : Any = {int(__lowerCamelCase ): v for k, v in idalabel.items()} __snake_case : Tuple = idalabel __snake_case : int = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: __snake_case : Tuple = 3_8_4 __snake_case : str = 1_5_3_6 __snake_case : Any = 6 elif "l16" in checkpoint_url: __snake_case : Any = 1_0_2_4 __snake_case : List[str] = 4_0_9_6 __snake_case : Optional[int] = 2_4 __snake_case : Dict = 1_6 __snake_case : Optional[int] = 0.1 elif "b4" in checkpoint_url: __snake_case : Tuple = 4 elif "l7" in checkpoint_url: __snake_case : List[Any] = 7 __snake_case : Dict = 1_0_2_4 __snake_case : Union[str, Any] = 4_0_9_6 __snake_case : Tuple = 2_4 __snake_case : Any = 1_6 __snake_case : Dict = 0.1 __snake_case : List[str] = ViTMSNModel(__lowerCamelCase ) __snake_case : str = torch.hub.load_state_dict_from_url(__lowerCamelCase , map_location="cpu" )["target_encoder"] __snake_case : Optional[int] = ViTImageProcessor(size=config.image_size ) remove_projection_head(__lowerCamelCase ) __snake_case : Tuple = create_rename_keys(__lowerCamelCase , base_model=__lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase , __lowerCamelCase , base_model=__lowerCamelCase ) model.load_state_dict(__lowerCamelCase ) model.eval() __snake_case : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" __snake_case : Tuple = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) __snake_case : str = ViTImageProcessor( size=config.image_size , image_mean=__lowerCamelCase , image_std=__lowerCamelCase ) __snake_case : str = image_processor(images=__lowerCamelCase , return_tensors="pt" ) # forward pass torch.manual_seed(2 ) __snake_case : str = model(**__lowerCamelCase ) __snake_case : Tuple = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: __snake_case : Any = torch.tensor([[-1.0_9_1_5, -1.4_8_7_6, -1.1_8_0_9]] ) elif "b16" in checkpoint_url: __snake_case : Any = torch.tensor([[1_4.2_8_8_9, -1_8.9_0_4_5, 1_1.7_2_8_1]] ) elif "l16" in checkpoint_url: __snake_case : Any = torch.tensor([[4_1.5_0_2_8, -2_2.8_6_8_1, 4_5.6_4_7_5]] ) elif "b4" in checkpoint_url: __snake_case : Optional[int] = torch.tensor([[-4.3_8_6_8, 5.2_9_3_2, -0.4_1_3_7]] ) else: __snake_case : List[Any] = torch.tensor([[-0.1_7_9_2, -0.6_4_6_5, 2.4_2_6_3]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , __lowerCamelCase , atol=1e-4 ) print(F'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _snake_case : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar", type=str, help="URL of the checkpoint you\'d like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) _snake_case : List[str] = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
123
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase = { '''configuration_bert''': ['''BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BertConfig''', '''BertOnnxConfig'''], '''tokenization_bert''': ['''BasicTokenizer''', '''BertTokenizer''', '''WordpieceTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''BertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BertForMaskedLM''', '''BertForMultipleChoice''', '''BertForNextSentencePrediction''', '''BertForPreTraining''', '''BertForQuestionAnswering''', '''BertForSequenceClassification''', '''BertForTokenClassification''', '''BertLayer''', '''BertLMHeadModel''', '''BertModel''', '''BertPreTrainedModel''', '''load_tf_weights_in_bert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBertEmbeddings''', '''TFBertForMaskedLM''', '''TFBertForMultipleChoice''', '''TFBertForNextSentencePrediction''', '''TFBertForPreTraining''', '''TFBertForQuestionAnswering''', '''TFBertForSequenceClassification''', '''TFBertForTokenClassification''', '''TFBertLMHeadModel''', '''TFBertMainLayer''', '''TFBertModel''', '''TFBertPreTrainedModel''', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFBertTokenizer'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxBertForCausalLM''', '''FlaxBertForMaskedLM''', '''FlaxBertForMultipleChoice''', '''FlaxBertForNextSentencePrediction''', '''FlaxBertForPreTraining''', '''FlaxBertForQuestionAnswering''', '''FlaxBertForSequenceClassification''', '''FlaxBertForTokenClassification''', '''FlaxBertModel''', '''FlaxBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ : Optional[Any] = { 'configuration_lxmert': ['LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LxmertConfig'], 'tokenization_lxmert': ['LxmertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Optional[int] = ['LxmertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Optional[Any] = [ 'LxmertEncoder', 'LxmertForPreTraining', 'LxmertForQuestionAnswering', 'LxmertModel', 'LxmertPreTrainedModel', 'LxmertVisualFeatureEncoder', 'LxmertXLayer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = [ 'TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLxmertForPreTraining', 'TFLxmertMainLayer', 'TFLxmertModel', 'TFLxmertPreTrainedModel', 'TFLxmertVisualFeatureEncoder', ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
48
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.linear_k''': '''encoder.layers.*.self_attn.linear_k''', '''self_attn.linear_v''': '''encoder.layers.*.self_attn.linear_v''', '''self_attn.linear_q''': '''encoder.layers.*.self_attn.linear_q''', '''self_attn.pos_bias_u''': '''encoder.layers.*.self_attn.pos_bias_u''', '''self_attn.pos_bias_v''': '''encoder.layers.*.self_attn.pos_bias_v''', '''self_attn.linear_out''': '''encoder.layers.*.self_attn.linear_out''', '''self_attn.linear_pos''': '''encoder.layers.*.self_attn.linear_pos''', '''self_attn.rotary_emb''': '''encoder.embed_positions''', '''self_attn_layer_norm''': '''encoder.layers.*.self_attn_layer_norm''', '''conv_module.pointwise_conv1''': '''encoder.layers.*.conv_module.pointwise_conv1''', '''conv_module.pointwise_conv2''': '''encoder.layers.*.conv_module.pointwise_conv2''', '''conv_module.depthwise_conv''': '''encoder.layers.*.conv_module.depthwise_conv''', '''conv_module.batch_norm''': '''encoder.layers.*.conv_module.batch_norm''', '''conv_module.layer_norm''': '''encoder.layers.*.conv_module.layer_norm''', '''ffn1.w_1''': '''encoder.layers.*.ffn1.intermediate_dense''', '''ffn1.w_2''': '''encoder.layers.*.ffn1.output_dense''', '''ffn1.layer_norm''': '''encoder.layers.*.ffn1_layer_norm''', '''ffn2.w_1''': '''encoder.layers.*.ffn2.intermediate_dense''', '''ffn2.w_2''': '''encoder.layers.*.ffn2.output_dense''', '''ffn2.layer_norm''': '''encoder.layers.*.ffn2_layer_norm''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __UpperCamelCase = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: for attribute in key.split('.' ): snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ) if weight_type is not None: snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ).shape else: snake_case_ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": snake_case_ = value elif weight_type == "weight_g": snake_case_ = value elif weight_type == "weight_v": snake_case_ = value elif weight_type == "bias": snake_case_ = value elif weight_type == "running_mean": snake_case_ = value elif weight_type == "running_var": snake_case_ = value elif weight_type == "num_batches_tracked": snake_case_ = value elif weight_type == "inv_freq": snake_case_ = value else: snake_case_ = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = [] snake_case_ = fairseq_model.state_dict() snake_case_ = hf_model.wavaveca_conformer.feature_extractor for name, value in fairseq_dict.items(): snake_case_ = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) snake_case_ = True else: for key, mapped_key in MAPPING.items(): snake_case_ = 'wav2vec2_conformer.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case_ = True if "*" in mapped_key: snake_case_ = name.split(UpperCAmelCase )[0].split('.' )[-2] snake_case_ = mapped_key.replace('*' , UpperCAmelCase ) if "pos_bias_u" in name: snake_case_ = None elif "pos_bias_v" in name: snake_case_ = None elif "weight_g" in name: snake_case_ = 'weight_g' elif "weight_v" in name: snake_case_ = 'weight_v' elif "bias" in name: snake_case_ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case_ = 'weight' elif "running_mean" in name: snake_case_ = 'running_mean' elif "inv_freq" in name: snake_case_ = 'inv_freq' elif "running_var" in name: snake_case_ = 'running_var' elif "num_batches_tracked" in name: snake_case_ = 'num_batches_tracked' else: snake_case_ = None set_recursively(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) continue if not is_used: unused_weights.append(UpperCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: snake_case_ = full_name.split('conv_layers.' )[-1] snake_case_ = name.split('.' ) snake_case_ = int(items[0] ) snake_case_ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase ) @torch.no_grad() def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=True ) -> str: if config_path is not None: snake_case_ = WavaVecaConformerConfig.from_pretrained(UpperCAmelCase , hidden_act='swish' ) else: snake_case_ = WavaVecaConformerConfig() if "rope" in checkpoint_path: snake_case_ = 'rotary' if is_finetuned: if dict_path: snake_case_ = Dictionary.load(UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq snake_case_ = target_dict.pad_index snake_case_ = target_dict.bos_index snake_case_ = target_dict.eos_index snake_case_ = len(target_dict.symbols ) snake_case_ = os.path.join(UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase ) ) return os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) snake_case_ = target_dict.indices # fairseq has the <pad> and <s> switched snake_case_ = 0 snake_case_ = 1 with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(UpperCAmelCase , UpperCAmelCase ) snake_case_ = WavaVecaCTCTokenizer( UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase , ) snake_case_ = True if config.feat_extract_norm == 'layer' else False snake_case_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=UpperCAmelCase , return_attention_mask=UpperCAmelCase , ) snake_case_ = WavaVecaProcessor(feature_extractor=UpperCAmelCase , tokenizer=UpperCAmelCase ) processor.save_pretrained(UpperCAmelCase ) snake_case_ = WavaVecaConformerForCTC(UpperCAmelCase ) else: snake_case_ = WavaVecaConformerForPreTraining(UpperCAmelCase ) if is_finetuned: snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: snake_case_ = argparse.Namespace(task='audio_pretraining' ) snake_case_ = fairseq.tasks.setup_task(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=UpperCAmelCase ) snake_case_ = model[0].eval() recursively_load_weights(UpperCAmelCase , UpperCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __UpperCamelCase = parser.parse_args() convert_wavaveca_conformer_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
69
0
def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = len(_UpperCAmelCase ), len(grid[0] ) if ( min(_UpperCAmelCase , _UpperCAmelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) SCREAMING_SNAKE_CASE_: List[Any] = 0 count += depth_first_search(_UpperCAmelCase , row + 1 , _UpperCAmelCase , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , row - 1 , _UpperCAmelCase , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , _UpperCAmelCase , col + 1 , _UpperCAmelCase ) count += depth_first_search(_UpperCAmelCase , _UpperCAmelCase , col - 1 , _UpperCAmelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase ) -> list: if len(UpperCAmelCase ) <= 1: return [tuple(UpperCAmelCase )] snake_case_ = [] def generate(UpperCAmelCase , UpperCAmelCase ): snake_case_ = [0] * n res.append(tuple(UpperCAmelCase ) ) snake_case_ = 0 while i < n: if c[i] < i: if i % 2 == 0: snake_case_ , snake_case_ = arr[i], arr[0] else: snake_case_ , snake_case_ = arr[i], arr[c[i]] res.append(tuple(UpperCAmelCase ) ) c[i] += 1 snake_case_ = 0 else: snake_case_ = 0 i += 1 generate(len(UpperCAmelCase ) , UpperCAmelCase ) return res if __name__ == "__main__": __UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
69
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """asapp/sew-tiny-100k""": """https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json""", # See all SEW models at https://huggingface.co/models?filter=sew } class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): __lowerCamelCase : List[str] ='sew' def __init__( self : Optional[Any] , __lowercase : Optional[int]=32 , __lowercase : Dict=768 , __lowercase : int=12 , __lowercase : List[str]=12 , __lowercase : Dict=3072 , __lowercase : Tuple=2 , __lowercase : Tuple="gelu" , __lowercase : Tuple=0.1 , __lowercase : int=0.1 , __lowercase : Union[str, Any]=0.1 , __lowercase : List[str]=0.0 , __lowercase : Any=0.1 , __lowercase : int=0.1 , __lowercase : int=0.02 , __lowercase : Any=1E-5 , __lowercase : int="group" , __lowercase : Union[str, Any]="gelu" , __lowercase : int=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , __lowercase : int=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , __lowercase : Tuple=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , __lowercase : Union[str, Any]=False , __lowercase : Union[str, Any]=128 , __lowercase : Optional[Any]=16 , __lowercase : Optional[Any]=True , __lowercase : Optional[int]=0.05 , __lowercase : Any=10 , __lowercase : str=2 , __lowercase : Union[str, Any]=0.0 , __lowercase : Union[str, Any]=10 , __lowercase : List[Any]=0 , __lowercase : Tuple="mean" , __lowercase : Any=False , __lowercase : Tuple=False , __lowercase : int=256 , __lowercase : Dict=0 , __lowercase : List[str]=1 , __lowercase : Optional[int]=2 , **__lowercase : str , ): '''simple docstring''' super().__init__(**lowerCAmelCase__ , pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) __a = hidden_size __a = feat_extract_norm __a = feat_extract_activation __a = list(lowerCAmelCase__ ) __a = list(lowerCAmelCase__ ) __a = list(lowerCAmelCase__ ) __a = conv_bias __a = num_conv_pos_embeddings __a = num_conv_pos_embedding_groups __a = len(self.conv_dim ) __a = num_hidden_layers __a = intermediate_size __a = squeeze_factor __a = hidden_act __a = num_attention_heads __a = hidden_dropout __a = attention_dropout __a = activation_dropout __a = feat_proj_dropout __a = final_dropout __a = layerdrop __a = layer_norm_eps __a = initializer_range __a = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __a = apply_spec_augment __a = mask_time_prob __a = mask_time_length __a = mask_time_min_masks __a = mask_feature_prob __a = mask_feature_length __a = mask_feature_min_masks # ctc loss __a = ctc_loss_reduction __a = ctc_zero_infinity # sequence classification __a = use_weighted_layer_sum __a = classifier_proj_size @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __UpperCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __UpperCamelCase = TaTokenizerFast __UpperCamelCase = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __UpperCamelCase = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
69
0
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 A__: Dict = data_utils.TransfoXLTokenizer A__: str = data_utils.TransfoXLCorpus A__: str = data_utils A__: Optional[int] = data_utils def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Optional[int] ,_UpperCAmelCase : str ,_UpperCAmelCase : List[Any] ,_UpperCAmelCase : str ) -> List[str]: if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(_UpperCAmelCase ,"""rb""" ) as fp: _a : int =pickle.load(_UpperCAmelCase ,encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) _a : Any =pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(F"Save vocabulary to {pytorch_vocab_dump_path}" ) _a : Union[str, Any] =corpus.vocab.__dict__ torch.save(_UpperCAmelCase ,_UpperCAmelCase ) _a : List[Any] =corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" ,_UpperCAmelCase ) _a : Any =pytorch_dump_folder_path + """/""" + CORPUS_NAME print(F"Save dataset to {pytorch_dataset_dump_path}" ) torch.save(_UpperCAmelCase ,_UpperCAmelCase ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model _a : int =os.path.abspath(_UpperCAmelCase ) _a : List[str] =os.path.abspath(_UpperCAmelCase ) print(F"Converting Transformer XL checkpoint from {tf_path} with config at {config_path}." ) # Initialise PyTorch model if transfo_xl_config_file == "": _a : Dict =TransfoXLConfig() else: _a : Any =TransfoXLConfig.from_json_file(_UpperCAmelCase ) print(F"Building PyTorch model from configuration: {config}" ) _a : Tuple =TransfoXLLMHeadModel(_UpperCAmelCase ) _a : Tuple =load_tf_weights_in_transfo_xl(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) # Save pytorch-model _a : str =os.path.join(_UpperCAmelCase ,_UpperCAmelCase ) _a : Optional[Any] =os.path.join(_UpperCAmelCase ,_UpperCAmelCase ) print(F"Save PyTorch model to {os.path.abspath(_UpperCAmelCase )}" ) torch.save(model.state_dict() ,_UpperCAmelCase ) print(F"Save configuration file to {os.path.abspath(_UpperCAmelCase )}" ) with open(_UpperCAmelCase ,"""w""" ,encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A__: Any = argparse.ArgumentParser() parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--tf_checkpoint_path''', default='''''', type=str, help='''An optional path to a TensorFlow checkpoint path to be converted.''', ) parser.add_argument( '''--transfo_xl_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--transfo_xl_dataset_file''', default='''''', type=str, help='''An optional dataset file to be converted in a vocabulary.''', ) A__: List[str] = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
276
"""simple docstring""" import sys from collections import defaultdict class UpperCamelCase : def __init__( self) -> Optional[int]: snake_case_ = [] def a_ ( self, lowerCAmelCase__) -> Any: return self.node_position[vertex] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = pos def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> str: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: snake_case_ = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: snake_case_ = 2 * start + 1 else: snake_case_ = 2 * start + 2 if heap[smallest_child] < heap[start]: snake_case_ , snake_case_ = heap[smallest_child], positions[smallest_child] snake_case_ , snake_case_ = ( heap[start], positions[start], ) snake_case_ , snake_case_ = temp, tempa snake_case_ = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child], self.get_position(positions[start])) self.set_position(positions[start], lowerCAmelCase__) self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> List[str]: snake_case_ = position[index] while index != 0: snake_case_ = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: snake_case_ = heap[parent] snake_case_ = position[parent] self.set_position(position[parent], lowerCAmelCase__) else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, lowerCAmelCase__) break snake_case_ = parent else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, 0) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = len(lowerCAmelCase__) // 2 - 1 for i in range(lowerCAmelCase__, -1, -1): self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, len(lowerCAmelCase__), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = positions[0] snake_case_ = sys.maxsize self.top_to_bottom(lowerCAmelCase__, 0, len(lowerCAmelCase__), lowerCAmelCase__) return temp def UpperCAmelCase ( UpperCAmelCase ) -> Tuple: snake_case_ = Heap() snake_case_ = [0] * len(UpperCAmelCase ) snake_case_ = [-1] * len(UpperCAmelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph snake_case_ = [] # Heap of Distance of vertices from their neighboring vertex snake_case_ = [] for vertex in range(len(UpperCAmelCase ) ): distance_tv.append(sys.maxsize ) positions.append(UpperCAmelCase ) heap.node_position.append(UpperCAmelCase ) snake_case_ = [] snake_case_ = 1 snake_case_ = sys.maxsize for neighbor, distance in adjacency_list[0]: snake_case_ = 0 snake_case_ = distance heap.heapify(UpperCAmelCase , UpperCAmelCase ) for _ in range(1 , len(UpperCAmelCase ) ): snake_case_ = heap.delete_minimum(UpperCAmelCase , UpperCAmelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) snake_case_ = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(UpperCAmelCase )] ): snake_case_ = distance heap.bottom_to_top( UpperCAmelCase , heap.get_position(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase ) snake_case_ = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > __UpperCamelCase = int(input('''Enter number of edges: ''').strip()) __UpperCamelCase = defaultdict(list) for _ in range(edges_number): __UpperCamelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
69
0
'''simple docstring''' import torch from transformers import AutoModel class a__ ( torch.nn.Module ): def __init__( self , _UpperCamelCase="sayef/fsner-bert-base-uncased" ): """simple docstring""" super(lowerCAmelCase__ , self ).__init__() _lowercase : Optional[Any] = AutoModel.from_pretrained(lowerCAmelCase__ , return_dict=lowerCAmelCase__ ) _lowercase : Optional[int] = torch.nn.CosineSimilarity(3 , 1E-0_8 ) _lowercase : Dict = torch.nn.Softmax(dim=1 ) def _lowerCamelCase ( self , **_UpperCamelCase ): """simple docstring""" return self.bert(**lowerCAmelCase__ ).last_hidden_state def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" return token_embeddings.sum(2 , keepdim=lowerCAmelCase__ ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=1 ): """simple docstring""" return self.softmax(T * self.cos(lowerCAmelCase__ , lowerCAmelCase__ ) ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : Dict = W_supports["sizes"].tolist() _lowercase : Tuple = W_supports["start_token_id"].item() _lowercase : Union[str, Any] = W_supports["end_token_id"].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] _lowercase : str = self.BERT(**lowerCAmelCase__ ) _lowercase : int = self.BERT(**lowerCAmelCase__ ) _lowercase : List[Any] = None _lowercase : int = None _lowercase : int = W_supports["input_ids"] == start_token_id _lowercase : Optional[Any] = W_supports["input_ids"] == end_token_id for i, size in enumerate(lowerCAmelCase__ ): if i == 0: _lowercase : Dict = 0 else: _lowercase : Optional[Any] = support_sizes[i - 1] _lowercase : List[str] = S[s : s + size][start_token_masks[s : s + size]] _lowercase : int = S[s : s + size][end_token_masks[s : s + size]] _lowercase : List[str] = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) _lowercase : Tuple = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: _lowercase : Dict = torch.vstack((p_starts, p_start) ) _lowercase : int = torch.vstack((p_ends, p_end) ) else: _lowercase : int = p_start _lowercase : int = p_end return p_starts, p_ends
250
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase = { '''configuration_mctct''': ['''MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MCTCTConfig'''], '''feature_extraction_mctct''': ['''MCTCTFeatureExtractor'''], '''processing_mctct''': ['''MCTCTProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MCTCTForCTC''', '''MCTCTModel''', '''MCTCTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
'''simple docstring''' from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def __UpperCAmelCase ( ): _UpperCAmelCase : int = HfArgumentParser(a_ ) _UpperCAmelCase : Optional[int] = parser.parse_args_into_dataclasses()[0] _UpperCAmelCase : List[Any] = TensorFlowBenchmark(args=a_ ) try: _UpperCAmelCase : Optional[int] = parser.parse_args_into_dataclasses()[0] except ValueError as e: _UpperCAmelCase : Optional[int] = "Arg --no_{0} is no longer used, please use --no-{0} instead." _UpperCAmelCase : List[str] = " ".join(str(a_ ).split(" " )[:-1] ) _UpperCAmelCase : Tuple = "" _UpperCAmelCase : List[Any] = eval(str(a_ ).split(" " )[-1] ) _UpperCAmelCase : Any = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(a_ ) if len(a_ ) > 0: _UpperCAmelCase : str = full_error_msg + begin_error_msg + str(a_ ) raise ValueError(a_ ) benchmark.run() if __name__ == "__main__": main()
145
"""simple docstring""" from math import factorial def UpperCAmelCase ( UpperCAmelCase = 20 ) -> int: snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(UpperCAmelCase ) / (factorial(UpperCAmelCase ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: __UpperCamelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
69
0
def a__ ( snake_case ): """simple docstring""" return "".join([hex(snake_case )[2:].zfill(2 ).upper() for byte in list(snake_case )] ) def a__ ( snake_case ): """simple docstring""" # Check data validity, following RFC3548 # https://www.ietf.org/rfc/rfc3548.txt if (len(snake_case ) % 2) != 0: raise ValueError( '''Base16 encoded data is invalid:\nData does not have an even number of hex digits.''' ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(snake_case ) <= set('''0123456789ABCDEF''' ): raise ValueError( '''Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters.''' ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(snake_case ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
303
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
0
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__ : def __init__( self , _a , _a=13 , _a=30 , _a=2 , _a=3 , _a=True , _a=True , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=10 , _a=0.0_2 , _a=None , _a=2 , ): lowercase : Optional[Any] = parent lowercase : Optional[Any] = batch_size lowercase : Dict = image_size lowercase : Dict = patch_size lowercase : Dict = num_channels lowercase : int = is_training lowercase : List[str] = use_labels lowercase : Any = hidden_size lowercase : int = num_hidden_layers lowercase : Optional[Any] = num_attention_heads lowercase : List[Any] = intermediate_size lowercase : Optional[int] = hidden_act lowercase : Optional[int] = hidden_dropout_prob lowercase : Optional[int] = attention_probs_dropout_prob lowercase : Optional[Any] = type_sequence_label_size lowercase : Optional[Any] = initializer_range lowercase : List[Any] = scope lowercase : Any = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowercase : Any = (image_size // patch_size) ** 2 lowercase : str = num_patches + 1 def __magic_name__ ( self ): lowercase : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase : List[Any] = None if self.use_labels: lowercase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase : Union[str, Any] = self.get_config() return config, pixel_values, labels def __magic_name__ ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __magic_name__ ( self , _a , _a , _a ): lowercase : str = ViTModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowercase : str = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ ( self , _a , _a , _a ): lowercase : str = ViTForMaskedImageModeling(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowercase : int = model(lowerCAmelCase__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowercase : List[str] = 1 lowercase : str = ViTForMaskedImageModeling(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowercase : Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase : str = model(lowerCAmelCase__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __magic_name__ ( self , _a , _a , _a ): lowercase : List[Any] = self.type_sequence_label_size lowercase : Optional[int] = ViTForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowercase : List[str] = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowercase : Dict = 1 lowercase : Dict = ViTForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowercase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase : Tuple = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __magic_name__ ( self ): lowercase : Union[str, Any] = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ) : List[Any] = config_and_inputs lowercase : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a__ ( lowerCAmelCase__, lowerCAmelCase__, unittest.TestCase ): __lowerCAmelCase = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) __lowerCAmelCase = ( {"""feature-extraction""": ViTModel, """image-classification""": ViTForImageClassification} if is_torch_available() else {} ) __lowerCAmelCase = True __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def __magic_name__ ( self ): lowercase : str = ViTModelTester(self ) lowercase : int = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=37 ) def __magic_name__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def __magic_name__ ( self ): pass def __magic_name__ ( self ): lowercase , lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Dict = model_class(lowerCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase__ , nn.Linear ) ) def __magic_name__ ( self ): lowercase , lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Union[str, Any] = model_class(lowerCAmelCase__ ) lowercase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase : Tuple = [*signature.parameters.keys()] lowercase : Union[str, Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def __magic_name__ ( self ): lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def __magic_name__ ( self ): lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCAmelCase__ ) def __magic_name__ ( self ): lowercase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) @slow def __magic_name__ ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase : Any = ViTModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __magic_name__ ( ) -> List[Any]: lowercase : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a__ ( unittest.TestCase ): @cached_property def __magic_name__ ( self ): return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def __magic_name__ ( self ): lowercase : Union[str, Any] = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(lowerCAmelCase__ ) lowercase : str = self.default_image_processor lowercase : str = prepare_img() lowercase : List[Any] = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): lowercase : Optional[int] = model(**lowerCAmelCase__ ) # verify the logits lowercase : str = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) lowercase : Union[str, Any] = torch.tensor([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def __magic_name__ ( self ): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. lowercase : List[str] = ViTModel.from_pretrained("facebook/dino-vits8" ).to(lowerCAmelCase__ ) lowercase : Union[str, Any] = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) lowercase : List[str] = prepare_img() lowercase : List[Any] = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ) lowercase : Dict = inputs.pixel_values.to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): lowercase : Dict = model(lowerCAmelCase__ , interpolate_pos_encoding=lowerCAmelCase__ ) # verify the logits lowercase : Any = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , lowerCAmelCase__ ) lowercase : Any = torch.tensor( [[4.2_3_4_0, 4.3_9_0_6, -6.6_6_9_2], [4.5_4_6_3, 1.8_9_2_8, -6.7_2_5_7], [4.4_4_2_9, 0.8_4_9_6, -5.8_5_8_5]] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def __magic_name__ ( self ): lowercase : Tuple = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) lowercase : Dict = self.default_image_processor lowercase : Optional[int] = prepare_img() lowercase : Optional[Any] = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ) lowercase : Optional[Any] = inputs.pixel_values.to(lowerCAmelCase__ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowercase : Optional[Any] = model(lowerCAmelCase__ )
202
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = (DPMSolverSinglestepScheduler,) SCREAMING_SNAKE_CASE_ = (("num_inference_steps", 2_5),) def a_ ( self, **lowerCAmelCase__) -> int: snake_case_ = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf'), 'variance_type': None, } config.update(**lowerCAmelCase__) return config def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> List[Any]: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ , snake_case_ = sample, sample for t in range(lowerCAmelCase__, time_step + scheduler.config.solver_order + 1): snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self) -> Union[str, Any]: pass def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> int: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self, lowerCAmelCase__=None, **lowerCAmelCase__) -> Union[str, Any]: if scheduler is None: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample return sample def a_ ( self) -> List[Any]: snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = 50 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2574) < 1e-3 def a_ ( self) -> Dict: for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__) def a_ ( self) -> Optional[Any]: # make sure that iterating over schedulers with same config names gives same results # for defaults snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 snake_case_ = DEISMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverMultistepScheduler.from_config(scheduler.config) snake_case_ = UniPCMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverSinglestepScheduler.from_config(scheduler.config) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> str: self.check_over_configs(thresholding=lowerCAmelCase__) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCAmelCase__, prediction_type=lowerCAmelCase__, sample_max_value=lowerCAmelCase__, algorithm_type='dpmsolver++', solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, ) def a_ ( self) -> Tuple: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__) def a_ ( self) -> Optional[int]: for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) snake_case_ = self.full_loop( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) assert not torch.isnan(lowerCAmelCase__).any(), "Samples have nan numbers" def a_ ( self) -> Optional[Any]: self.check_over_configs(lower_order_final=lowerCAmelCase__) self.check_over_configs(lower_order_final=lowerCAmelCase__) def a_ ( self) -> Any: self.check_over_configs(lambda_min_clipped=-float('inf')) self.check_over_configs(lambda_min_clipped=-5.1) def a_ ( self) -> Any: self.check_over_configs(variance_type=lowerCAmelCase__) self.check_over_configs(variance_type='learned_range') def a_ ( self) -> List[Any]: for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=lowerCAmelCase__, time_step=0) def a_ ( self) -> int: snake_case_ = self.full_loop() snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> Dict: snake_case_ = self.full_loop(use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2248) < 1e-3 def a_ ( self) -> Union[str, Any]: snake_case_ = self.full_loop(prediction_type='v_prediction') snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.1453) < 1e-3 def a_ ( self) -> Optional[Any]: snake_case_ = self.full_loop(prediction_type='v_prediction', use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.0649) < 1e-3 def a_ ( self) -> Optional[int]: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(thresholding=lowerCAmelCase__, dynamic_thresholding_ratio=0) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample assert sample.dtype == torch.floataa
69
0
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING lowercase : Optional[Any] = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase__ ) class A__ ( lowerCAmelCase__ ): """simple docstring""" def __init__( self , *lowercase , **lowercase) -> Tuple: '''simple docstring''' super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__) self.check_model_type(lowerCAmelCase__) def __lowercase ( self , lowercase=None , lowercase=None , lowercase=None , **lowercase) -> Any: '''simple docstring''' a__ , a__ : str = {}, {} if padding is not None: a__ : Any = padding if truncation is not None: a__ : str = truncation if top_k is not None: a__ : Optional[Any] = top_k return preprocess_params, {}, postprocess_params def __call__( self , lowercase , lowercase = None , **lowercase) -> Optional[Any]: '''simple docstring''' if isinstance(lowerCAmelCase__ , (Image.Image, str)) and isinstance(lowerCAmelCase__ , lowerCAmelCase__): a__ : Dict = {'image': image, 'question': question} else: a__ : List[str] = image a__ : Optional[int] = super().__call__(lowerCAmelCase__ , **lowerCAmelCase__) return results def __lowercase ( self , lowercase , lowercase=False , lowercase=False) -> int: '''simple docstring''' a__ : Union[str, Any] = load_image(inputs['image']) a__ : Any = self.tokenizer( inputs['question'] , return_tensors=self.framework , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__) a__ : Optional[Any] = self.image_processor(images=lowerCAmelCase__ , return_tensors=self.framework) model_inputs.update(lowerCAmelCase__) return model_inputs def __lowercase ( self , lowercase) -> Optional[int]: '''simple docstring''' a__ : List[str] = self.model(**lowerCAmelCase__) return model_outputs def __lowercase ( self , lowercase , lowercase=5) -> Dict: '''simple docstring''' if top_k > self.model.config.num_labels: a__ : Optional[Any] = self.model.config.num_labels if self.framework == "pt": a__ : List[Any] = model_outputs.logits.sigmoid()[0] a__ , a__ : List[str] = probs.topk(lowerCAmelCase__) else: raise ValueError(F'Unsupported framework: {self.framework}') a__ : Dict = scores.tolist() a__ : str = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(lowerCAmelCase__ , lowerCAmelCase__)]
99
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # Base Case if curr_ind == len(UpperCAmelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(UpperCAmelCase ) ): if valid_connection(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # Insert current vertex into path as next transition snake_case_ = next_ver # Validate created path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , curr_ind + 1 ): return True # Backtrack snake_case_ = -1 return False def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = 0 ) -> list[int]: snake_case_ = [-1] * (len(UpperCAmelCase ) + 1) # initialize start and end of path with starting index snake_case_ = snake_case_ = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , 1 ) else []
69
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ : Optional[int] = logging.get_logger(__name__) A_ : str = { 'facebook/levit-128S': 'https://huggingface.co/facebook/levit-128S/resolve/main/config.json', # See all LeViT models at https://huggingface.co/models?filter=levit } class A_ ( lowerCAmelCase__ ): '''simple docstring''' a__ = "levit" def __init__(self , lowercase__=224 , lowercase__=3 , lowercase__=3 , lowercase__=2 , lowercase__=1 , lowercase__=16 , lowercase__=[128, 256, 384] , lowercase__=[4, 8, 12] , lowercase__=[4, 4, 4] , lowercase__=[16, 16, 16] , lowercase__=0 , lowercase__=[2, 2, 2] , lowercase__=[2, 2, 2] , lowercase__=0.02 , **lowercase__ , ) -> Optional[Any]: super().__init__(**lowerCAmelCase__ ) __UpperCAmelCase = image_size __UpperCAmelCase = num_channels __UpperCAmelCase = kernel_size __UpperCAmelCase = stride __UpperCAmelCase = padding __UpperCAmelCase = hidden_sizes __UpperCAmelCase = num_attention_heads __UpperCAmelCase = depths __UpperCAmelCase = key_dim __UpperCAmelCase = drop_path_rate __UpperCAmelCase = patch_size __UpperCAmelCase = attention_ratio __UpperCAmelCase = mlp_ratio __UpperCAmelCase = initializer_range __UpperCAmelCase = [ ['''Subsample''', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['''Subsample''', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] class A_ ( lowerCAmelCase__ ): '''simple docstring''' a__ = version.parse("1.11" ) @property def lowerCAmelCase_ (self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCAmelCase_ (self ) -> float: return 1E-4
333
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''▁''' __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model''' ), } } __UpperCamelCase = { '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__ = None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = {} if sp_model_kwargs is None else sp_model_kwargs snake_case_ = legacy_behaviour super().__init__( bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, sp_model_kwargs=self.sp_model_kwargs, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowerCAmelCase__)) snake_case_ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token snake_case_ = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab snake_case_ = 1 snake_case_ = len(self.sp_model) snake_case_ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase__) } snake_case_ = {v: k for k, v in self.lang_code_to_id.items()} snake_case_ = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) snake_case_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} snake_case_ = list(self.lang_code_to_id.keys()) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens]) snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.lang_code_to_id[self._src_lang] snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self) -> Union[str, Any]: snake_case_ = self.__dict__.copy() snake_case_ = None snake_case_ = self.sp_model.serialized_model_proto() return state def __setstate__( self, lowerCAmelCase__) -> Tuple: snake_case_ = d # for backward compatibility if not hasattr(self, 'sp_model_kwargs'): snake_case_ = {} snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) @property def a_ ( self) -> str: return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None, lowerCAmelCase__ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__, token_ids_a=lowerCAmelCase__, already_has_special_tokens=lowerCAmelCase__) snake_case_ = [1] * len(self.prefix_tokens) snake_case_ = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__)) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__)) + ([0] * len(lowerCAmelCase__)) + suffix_ones def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self) -> List[Any]: snake_case_ = {self.convert_ids_to_tokens(lowerCAmelCase__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def a_ ( self, lowerCAmelCase__) -> List[str]: return self.sp_model.encode(lowerCAmelCase__, out_type=lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> Any: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case_ = self.sp_model.PieceToId(lowerCAmelCase__) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a_ ( self, lowerCAmelCase__) -> Dict: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def a_ ( self, lowerCAmelCase__) -> List[str]: snake_case_ = ''.join(lowerCAmelCase__).replace(lowerCAmelCase__, ' ').strip() return out_string def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file, lowerCAmelCase__) elif not os.path.isfile(self.vocab_file): with open(lowerCAmelCase__, 'wb') as fi: snake_case_ = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__) return (out_vocab_file,) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> Union[str, Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> int: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[src_lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id]
69
0
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder _snake_case : Tuple = datasets.utils.logging.get_logger(__name__) class a (folder_based_builder.FolderBasedBuilderConfig ): """simple docstring""" __UpperCAmelCase : int = None __UpperCAmelCase : Any = None class a (folder_based_builder.FolderBasedBuilder ): """simple docstring""" __UpperCAmelCase : Tuple = datasets.Audio() __UpperCAmelCase : List[str] = "audio" __UpperCAmelCase : List[Any] = AudioFolderConfig __UpperCAmelCase : Optional[int] = 42 # definition at the bottom of the script __UpperCAmelCase : Optional[int] = AudioClassification(audio_column="audio" , label_column="label" ) _snake_case : Optional[int] = [ ".aiff", ".au", ".avr", ".caf", ".flac", ".htk", ".svx", ".mat4", ".mat5", ".mpc2k", ".ogg", ".paf", ".pvf", ".raw", ".rf64", ".sd2", ".sds", ".ircam", ".voc", ".w64", ".wav", ".nist", ".wavex", ".wve", ".xi", ".mp3", ".opus", ] _snake_case : Any = AUDIO_EXTENSIONS
123
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCAmelCase ( ) -> int: snake_case_ = HfArgumentParser(UpperCAmelCase ) snake_case_ = parser.parse_args_into_dataclasses()[0] snake_case_ = TensorFlowBenchmark(args=UpperCAmelCase ) try: snake_case_ = parser.parse_args_into_dataclasses()[0] except ValueError as e: snake_case_ = 'Arg --no_{0} is no longer used, please use --no-{0} instead.' snake_case_ = ' '.join(str(UpperCAmelCase ).split(' ' )[:-1] ) snake_case_ = '' snake_case_ = eval(str(UpperCAmelCase ).split(' ' )[-1] ) snake_case_ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = full_error_msg + begin_error_msg + str(UpperCAmelCase ) raise ValueError(UpperCAmelCase ) benchmark.run() if __name__ == "__main__": main()
69
0
import argparse import os import torch from transformers.utils import WEIGHTS_NAME SCREAMING_SNAKE_CASE__ : Optional[int] = ['small', 'medium', 'large'] SCREAMING_SNAKE_CASE__ : Any = 'lm_head.decoder.weight' SCREAMING_SNAKE_CASE__ : List[str] = 'lm_head.weight' def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[Any]: lowerCamelCase : Optional[Any] = torch.load(_SCREAMING_SNAKE_CASE ) lowerCamelCase : Optional[int] = d.pop(_SCREAMING_SNAKE_CASE ) os.makedirs(_SCREAMING_SNAKE_CASE ,exist_ok=_SCREAMING_SNAKE_CASE ) torch.save(_SCREAMING_SNAKE_CASE ,os.path.join(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Dict = argparse.ArgumentParser() parser.add_argument('--dialogpt_path', default='.', type=str) SCREAMING_SNAKE_CASE__ : Union[str, Any] = parser.parse_args() for MODEL in DIALOGPT_MODELS: SCREAMING_SNAKE_CASE__ : int = os.path.join(args.dialogpt_path, f'''{MODEL}_ft.pkl''') SCREAMING_SNAKE_CASE__ : Optional[int] = f'''./DialoGPT-{MODEL}''' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
48
"""simple docstring""" from __future__ import annotations def UpperCAmelCase ( UpperCAmelCase ) -> None: create_state_space_tree(UpperCAmelCase , [] , 0 , [0 for i in range(len(UpperCAmelCase ) )] ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) -> None: if index == len(UpperCAmelCase ): print(UpperCAmelCase ) return for i in range(len(UpperCAmelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) snake_case_ = True create_state_space_tree(UpperCAmelCase , UpperCAmelCase , index + 1 , UpperCAmelCase ) current_sequence.pop() snake_case_ = False __UpperCamelCase = [3, 1, 2, 4] generate_all_permutations(sequence) __UpperCamelCase = ["A", "B", "C"] generate_all_permutations(sequence_a)
69
0
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore lowerCAmelCase : List[Any] = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" lowerCAmelCase : Dict = [file for file in filepaths if file != file.lower()] if upper_files: print(f'''{len(upper_files)} files contain uppercase characters:''') print("""\n""".join(upper_files) + """\n""") lowerCAmelCase : Any = [file for file in filepaths if """ """ in file] if space_files: print(f'''{len(space_files)} files contain space characters:''') print("""\n""".join(space_files) + """\n""") lowerCAmelCase : str = [file for file in filepaths if """-""" in file] if hyphen_files: print(f'''{len(hyphen_files)} files contain hyphen characters:''') print("""\n""".join(hyphen_files) + """\n""") lowerCAmelCase : Tuple = [file for file in filepaths if os.sep not in file] if nodir_files: print(f'''{len(nodir_files)} files are not in a directory:''') print("""\n""".join(nodir_files) + """\n""") lowerCAmelCase : Dict = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( 'Warning: upper bound of deterministic test is exceeded. ' 'Pass allow_probable=True to allow probabilistic test. ' 'A return value of True indicates a probable prime.' ) # array bounds provided by analysis snake_case_ = [ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] snake_case_ = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(UpperCAmelCase , 1 ): if n < _p: # then we have our last prime to check snake_case_ = primes[:idx] break snake_case_ , snake_case_ = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: snake_case_ = False for r in range(UpperCAmelCase ): snake_case_ = pow(UpperCAmelCase , d * 2**r , UpperCAmelCase ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): snake_case_ = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def UpperCAmelCase ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
69
0
import functools from typing import Any def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or len(_SCREAMING_SNAKE_CASE ) == 0: raise ValueError("""the string should be not empty string""" ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not all( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and len(_SCREAMING_SNAKE_CASE ) > 0 for item in words ): raise ValueError("""the words should be a list of non-empty strings""" ) # Build trie __a = {} __a = """WORD_KEEPER""" for word in words: __a = trie for c in word: if c not in trie_node: __a = {} __a = trie_node[c] __a = True __a = len(_SCREAMING_SNAKE_CASE ) # Dynamic programming method @functools.cache def is_breakable(_SCREAMING_SNAKE_CASE : Dict ) -> bool: if index == len_string: return True __a = trie for i in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __a = trie_node.get(string[i] , _SCREAMING_SNAKE_CASE ) if trie_node is None: return False if trie_node.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __UpperCamelCase = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
69
0
'''simple docstring''' import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class A__ ( lowerCAmelCase__ ): __UpperCamelCase : Optional[int] = "facebook/bart-large-mnli" __UpperCamelCase : Optional[Any] = ( "This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which " "should be the text to classify, and `labels`, which should be the list of labels to use for classification. " "It returns the most likely label in the list of provided `labels` for the input text." ) __UpperCamelCase : str = "text_classifier" __UpperCamelCase : Optional[int] = AutoTokenizer __UpperCamelCase : Union[str, Any] = AutoModelForSequenceClassification __UpperCamelCase : List[str] = ["text", ["text"]] __UpperCamelCase : List[str] = ["text"] def __UpperCAmelCase ( self :List[Any] ) -> Dict: '''simple docstring''' super().setup() _a : Optional[Any] =self.model.config _a : Optional[int] =-1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): _a : Optional[int] =int(lowerCAmelCase__ ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def __UpperCAmelCase ( self :int , SCREAMING_SNAKE_CASE :Optional[int] , SCREAMING_SNAKE_CASE :Optional[Any] ) -> List[Any]: '''simple docstring''' _a : Optional[int] =labels return self.pre_processor( [text] * len(lowerCAmelCase__ ) , [f"This example is {label}" for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def __UpperCAmelCase ( self :int , SCREAMING_SNAKE_CASE :Any ) -> Tuple: '''simple docstring''' _a : int =outputs.logits _a : str =torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
276
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: __UpperCamelCase = None __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } __UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = NllbTokenizer SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True snake_case_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) snake_case_ = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__) for lang_code in FAIRSEQ_LANGUAGE_CODES } snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.convert_tokens_to_ids(self._src_lang) snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> Tuple: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__): copyfile(self.vocab_file, lowerCAmelCase__) return (out_vocab_file,)
69
0
'''simple docstring''' from math import factorial def _A ( snake_case = 20 ) -> int: _lowercase : Any = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... _lowercase : Dict = n // 2 return int(factorial(snake_case ) / (factorial(snake_case ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: _snake_case = int(sys.argv[1]) print(solution(n)) except ValueError: print('Invalid entry - please enter a number.')
250
"""simple docstring""" from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''T5Config''' class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig
69
0
'''simple docstring''' from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass __a = (3, 9, -11, 0, 7, 5, 1, -1) __a = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class A__ : """simple docstring""" UpperCamelCase_ : Optional[Any] = 42 UpperCamelCase_ : Optional[int] = 42 class A__ : """simple docstring""" def __init__( self : Any , lowerCAmelCase__ : Dict ) -> None: """simple docstring""" _UpperCAmelCase : int = None for i in sorted(lowerCAmelCase__ , reverse=lowerCAmelCase__ ): _UpperCAmelCase : List[str] = Node(lowerCAmelCase__ , self.head ) def __iter__( self : Optional[int] ) -> Iterator[int]: """simple docstring""" _UpperCAmelCase : Union[str, Any] = self.head while node: yield node.data _UpperCAmelCase : Union[str, Any] = node.next_node def __len__( self : Dict ) -> int: """simple docstring""" return sum(1 for _ in self ) def __str__( self : Union[str, Any] ) -> str: """simple docstring""" return " -> ".join([str(lowerCAmelCase__ ) for node in self] ) def __UpperCAmelCase ( a_: List[str], a_: Any ): return SortedLinkedList(list(a_ ) + list(a_ ) ) if __name__ == "__main__": import doctest doctest.testmod() __a = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
145
"""simple docstring""" import argparse __UpperCamelCase = '''docs/source/_static/js/custom.js''' def UpperCAmelCase ( UpperCAmelCase ) -> int: with open(UpperCAmelCase , encoding='utf-8' , newline='\n' ) as f: snake_case_ = f.readlines() snake_case_ = 0 # First let's put the right version while not lines[index].startswith('const stableVersion =' ): index += 1 snake_case_ = f'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith('const versionMapping = {' ): index += 1 # We go until the end while not lines[index].startswith('}' ): index += 1 # We add the new version at the end lines[index - 1] += f' "v{version}": "v{version}",\n' with open(UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') __UpperCamelCase = parser.parse_args() update_custom_js(args.version)
69
0
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowercase_ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowercase_ = """cuda""" if torch.cuda.is_available() else """cpu""" def a__ ( snake_case , snake_case=100 , snake_case=" " ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = text.split(snake_case ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(snake_case ) , snake_case )] def a__ ( snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Union[str, Any] = [], [] for title, text in zip(documents['''title'''] , documents['''text'''] ): if text is not None: for passage in split_text(snake_case ): titles.append(title if title is not None else '''''' ) texts.append(snake_case ) return {"title": titles, "text": texts} def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = ctx_tokenizer( documents['''title'''] , documents['''text'''] , truncation=snake_case , padding='''longest''' , return_tensors='''pt''' )['''input_ids'''] __SCREAMING_SNAKE_CASE : List[Any] = ctx_encoder(input_ids.to(device=snake_case ) , return_dict=snake_case ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def a__ ( snake_case , snake_case , snake_case , ): """simple docstring""" ###################################### logger.info('''Step 1 - Create the dataset''' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way __SCREAMING_SNAKE_CASE : List[str] = load_dataset( '''csv''' , data_files=[rag_example_args.csv_path] , split='''train''' , delimiter='''\t''' , column_names=['''title''', '''text'''] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words __SCREAMING_SNAKE_CASE : Optional[Any] = dataset.map(snake_case , batched=snake_case , num_proc=processing_args.num_proc ) # And compute the embeddings __SCREAMING_SNAKE_CASE : Union[str, Any] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=snake_case ) __SCREAMING_SNAKE_CASE : List[str] = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) __SCREAMING_SNAKE_CASE : List[Any] = Features( {'''text''': Value('''string''' ), '''title''': Value('''string''' ), '''embeddings''': Sequence(Value('''float32''' ) )} ) # optional, save as float32 instead of float64 to save space __SCREAMING_SNAKE_CASE : Optional[int] = dataset.map( partial(snake_case , ctx_encoder=snake_case , ctx_tokenizer=snake_case ) , batched=snake_case , batch_size=processing_args.batch_size , features=snake_case , ) # And finally save your dataset __SCREAMING_SNAKE_CASE : str = os.path.join(rag_example_args.output_dir , '''my_knowledge_dataset''' ) dataset.save_to_disk(snake_case ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('''Step 2 - Index the dataset''' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search __SCREAMING_SNAKE_CASE : Union[str, Any] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('''embeddings''' , custom_index=snake_case ) # And save the index __SCREAMING_SNAKE_CASE : Tuple = os.path.join(rag_example_args.output_dir , '''my_knowledge_dataset_hnsw_index.faiss''' ) dataset.get_index('''embeddings''' ).save(snake_case ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class __UpperCamelCase : """simple docstring""" lowerCAmelCase_ = field( default=str(Path(lowerCAmelCase__ ).parent / '''test_run''' / '''dummy-kb''' / '''my_knowledge_dataset.csv''' ) , metadata={'''help''': '''Path to a tab-separated csv file with columns \'title\' and \'text\''''} , ) lowerCAmelCase_ = field( default=lowerCAmelCase__ , metadata={'''help''': '''Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'''} , ) lowerCAmelCase_ = field( default='''facebook/rag-sequence-nq''' , metadata={'''help''': '''The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''''} , ) lowerCAmelCase_ = field( default='''facebook/dpr-ctx_encoder-multiset-base''' , metadata={ '''help''': ( '''The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or''' ''' \'facebook/dpr-ctx_encoder-multiset-base\'''' ) } , ) lowerCAmelCase_ = field( default=str(Path(lowerCAmelCase__ ).parent / '''test_run''' / '''dummy-kb''' ) , metadata={'''help''': '''Path to a directory where the dataset passages and the index will be saved'''} , ) @dataclass class __UpperCamelCase : """simple docstring""" lowerCAmelCase_ = field( default=lowerCAmelCase__ , metadata={ '''help''': '''The number of processes to use to split the documents into passages. Default is single process.''' } , ) lowerCAmelCase_ = field( default=16 , metadata={ '''help''': '''The batch size to use when computing the passages embeddings using the DPR context encoder.''' } , ) @dataclass class __UpperCamelCase : """simple docstring""" lowerCAmelCase_ = field( default=7_68 , metadata={'''help''': '''The dimension of the embeddings to pass to the HNSW Faiss index.'''} , ) lowerCAmelCase_ = field( default=1_28 , metadata={ '''help''': ( '''The number of bi-directional links created for every new element during the HNSW index construction.''' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowercase_ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowercase_ , lowercase_ , lowercase_ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowercase_ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
303
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class UpperCamelCase : def __init__( self, lowerCAmelCase__) -> Optional[int]: snake_case_ = data snake_case_ = None class UpperCamelCase : def __init__( self) -> Dict: snake_case_ = None snake_case_ = None def __iter__( self) -> Iterator[Any]: snake_case_ = self.head while self.head: yield node.data snake_case_ = node.next if node == self.head: break def __len__( self) -> int: return sum(1 for _ in self) def __repr__( self) -> str: return "->".join(str(lowerCAmelCase__) for item in iter(self)) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(len(self), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(0, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> None: if index < 0 or index > len(self): raise IndexError('list index out of range.') snake_case_ = Node(lowerCAmelCase__) if self.head is None: snake_case_ = new_node # first node points itself snake_case_ = snake_case_ = new_node elif index == 0: # insert at head snake_case_ = self.head snake_case_ = snake_case_ = new_node else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = new_node if index == len(self) - 1: # insert at tail snake_case_ = new_node def a_ ( self) -> str: return self.delete_nth(0) def a_ ( self) -> Any: return self.delete_nth(len(self) - 1) def a_ ( self, lowerCAmelCase__ = 0) -> Any: if not 0 <= index < len(self): raise IndexError('list index out of range.') snake_case_ = self.head if self.head == self.tail: # just one node snake_case_ = snake_case_ = None elif index == 0: # delete head node snake_case_ = self.tail.next.next snake_case_ = self.head.next else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = temp.next.next if index == len(self) - 1: # delete at tail snake_case_ = temp return delete_node.data def a_ ( self) -> bool: return len(self) == 0 def UpperCAmelCase ( ) -> None: snake_case_ = CircularLinkedList() assert len(UpperCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(UpperCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(UpperCAmelCase ) == i circular_linked_list.insert_nth(UpperCAmelCase , i + 1 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
69
0
"""simple docstring""" import warnings from functools import wraps from typing import Callable def __magic_name__ ( __snake_case : Optional[int] ) -> Callable: @wraps(__snake_case ) def _inner_fn(*__snake_case : List[str] , **__snake_case : str ): warnings.warn( (f"""\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.""") , __snake_case , ) return fn(*__snake_case , **__snake_case ) return _inner_fn
202
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCamelCase = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __UpperCamelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) __UpperCamelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __UpperCamelCase = re.compile(r'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __UpperCamelCase = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def UpperCAmelCase ( UpperCAmelCase ) -> List[Any]: snake_case_ = None # source code of `config_class` snake_case_ = inspect.getsource(UpperCAmelCase ) snake_case_ = _re_checkpoint.findall(UpperCAmelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): snake_case_ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: snake_case_ = ckpt_name break return checkpoint def UpperCAmelCase ( ) -> Union[str, Any]: snake_case_ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ = get_checkpoint_from_config_class(UpperCAmelCase ) snake_case_ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = '\n'.join(sorted(UpperCAmelCase ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
from __future__ import annotations lowercase : Any = list[list[int]] # assigning initial values to the grid lowercase : List[Any] = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution lowercase : List[Any] = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def A_ ( A__ , A__ , A__ , A__ ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def A_ ( A__ ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def A_ ( A__ ) -> Matrix | None: if location := find_empty_location(A__ ): a__ , a__ : Union[str, Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(A__ , A__ , A__ , A__ ): a__ : Optional[Any] = digit if sudoku(A__ ) is not None: return grid a__ : Any = 0 return None def A_ ( A__ ) -> None: for row in grid: for cell in row: print(A__ , end=' ' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("""\nExample grid:\n""" + """=""" * 2_0) print_solution(example_grid) print("""\nExample grid solution:""") lowercase : Dict = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("""Cannot find a solution.""")
99
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __UpperCamelCase = 0 __UpperCamelCase = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __UpperCamelCase = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __UpperCamelCase = tuple[int, int] class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, ) -> None: snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = g_cost snake_case_ = parent snake_case_ = self.calculate_heuristic() snake_case_ = self.g_cost + self.h_cost def a_ ( self) -> float: snake_case_ = self.pos_x - self.goal_x snake_case_ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase__) + abs(lowerCAmelCase__) else: return sqrt(dy**2 + dx**2) def __lt__( self, lowerCAmelCase__) -> bool: return self.f_cost < other.f_cost class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = Node(start[1], start[0], goal[1], goal[0], 0, lowerCAmelCase__) snake_case_ = Node(goal[1], goal[0], goal[1], goal[0], 9_9999, lowerCAmelCase__) snake_case_ = [self.start] snake_case_ = [] snake_case_ = False def a_ ( self) -> list[TPosition]: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() snake_case_ = self.open_nodes.pop(0) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase__) self.closed_nodes.append(lowerCAmelCase__) snake_case_ = self.get_successors(lowerCAmelCase__) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__) else: self.open_nodes.append(lowerCAmelCase__) return [self.start.pos] def a_ ( self, lowerCAmelCase__) -> list[Node]: snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(lowerCAmelCase__) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__, lowerCAmelCase__, self.target.pos_y, self.target.pos_x, parent.g_cost + 1, lowerCAmelCase__, )) return successors def a_ ( self, lowerCAmelCase__) -> list[TPosition]: snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x)) snake_case_ = current_node.parent path.reverse() return path class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> None: snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = False def a_ ( self) -> list[TPosition]: while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() snake_case_ = self.fwd_astar.open_nodes.pop(0) snake_case_ = self.bwd_astar.open_nodes.pop(0) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase__, lowerCAmelCase__) self.fwd_astar.closed_nodes.append(lowerCAmelCase__) self.bwd_astar.closed_nodes.append(lowerCAmelCase__) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase__), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase__), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase__) else: astar.open_nodes.append(lowerCAmelCase__) return [self.fwd_astar.start.pos] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> list[TPosition]: snake_case_ = self.fwd_astar.retrace_path(lowerCAmelCase__) snake_case_ = self.bwd_astar.retrace_path(lowerCAmelCase__) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __UpperCamelCase = (0, 0) __UpperCamelCase = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __UpperCamelCase = time.time() __UpperCamelCase = AStar(init, goal) __UpperCamelCase = a_star.search() __UpperCamelCase = time.time() - start_time print(F"""AStar execution time = {end_time:f} seconds""") __UpperCamelCase = time.time() __UpperCamelCase = BidirectionalAStar(init, goal) __UpperCamelCase = time.time() - bd_start_time print(F"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
69
0
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ) -> Union[str, Any]: '''simple docstring''' # set parameter of one layer assert torch_layer.weight.shape == weight.shape, f'''{torch_layer} layer.weight does not match''' __UpperCAmelCase = nn.Parameter(SCREAMING_SNAKE_CASE ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f'''{torch_layer} layer.bias does not match''' __UpperCAmelCase = nn.Parameter(SCREAMING_SNAKE_CASE ) def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: '''simple docstring''' # set torch weights for 1-to-1 comparison __UpperCAmelCase = np.asarray(weights[0] ) __UpperCAmelCase = np.asarray(weights[1] ) __UpperCAmelCase = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE ) , ) set_param( torch_layer.self_attention.value , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE ) , ) set_param( torch_layer.output.dense , torch.tensor(SCREAMING_SNAKE_CASE ).view(-1 , SCREAMING_SNAKE_CASE ).contiguous().transpose(0 , 1 ) , ) def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' # set torch weights for 1-to-1 comparison __UpperCAmelCase = np.asarray(weights[0] ) __UpperCAmelCase = np.asarray(weights[1] ) __UpperCAmelCase = np.asarray(weights[2] ) __UpperCAmelCase = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE ) , ) set_param( torch_layer.self_attention.key , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE ) , ) set_param( torch_layer.self_attention.value , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE ) , ) set_param( torch_layer.output.dense , torch.tensor(SCREAMING_SNAKE_CASE ).view(-1 , SCREAMING_SNAKE_CASE ).contiguous().transpose(0 , 1 ) , ) def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' # layernorm 1 __UpperCAmelCase = weights[0][0][0] __UpperCAmelCase = np.asarray(layer_norm_a[0] ) __UpperCAmelCase = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(SCREAMING_SNAKE_CASE ) , torch.tensor(SCREAMING_SNAKE_CASE ) , ) # lsh weights + output __UpperCAmelCase = weights[0][1] if len(SCREAMING_SNAKE_CASE ) < 4: set_layer_weights_in_torch_lsh(SCREAMING_SNAKE_CASE , torch_block.attention , SCREAMING_SNAKE_CASE ) else: set_layer_weights_in_torch_local(SCREAMING_SNAKE_CASE , torch_block.attention , SCREAMING_SNAKE_CASE ) # intermediate weighs __UpperCAmelCase = weights[2][0][1][2] # Chunked Feed Forward if len(SCREAMING_SNAKE_CASE ) == 4: __UpperCAmelCase = intermediate_weights[2] # layernorm 2 __UpperCAmelCase = np.asarray(intermediate_weights[0][0] ) __UpperCAmelCase = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(SCREAMING_SNAKE_CASE ) , torch.tensor(SCREAMING_SNAKE_CASE ) , ) # intermediate dense __UpperCAmelCase = np.asarray(intermediate_weights[1][0] ) __UpperCAmelCase = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(0 , 1 ).contiguous() , torch.tensor(SCREAMING_SNAKE_CASE ) , ) # intermediate out __UpperCAmelCase = np.asarray(intermediate_weights[4][0] ) __UpperCAmelCase = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(0 , 1 ).contiguous() , torch.tensor(SCREAMING_SNAKE_CASE ) , ) def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[Any]: '''simple docstring''' # reformer model __UpperCAmelCase = torch_model.reformer # word embeds __UpperCAmelCase = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(SCREAMING_SNAKE_CASE ) , ) if isinstance(weights[3] , SCREAMING_SNAKE_CASE ): __UpperCAmelCase = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): __UpperCAmelCase = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f'''{position_embeddings[emb_idx]} emb does not match''' __UpperCAmelCase = nn.Parameter(torch.tensor(SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( SCREAMING_SNAKE_CASE ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): __UpperCAmelCase = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # output layer norm __UpperCAmelCase = np.asarray(weights[7][0] ) __UpperCAmelCase = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(SCREAMING_SNAKE_CASE ) , torch.tensor(SCREAMING_SNAKE_CASE ) , ) # output embeddings __UpperCAmelCase = np.asarray(weights[9][0] ) __UpperCAmelCase = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(SCREAMING_SNAKE_CASE ).transpose(0 , 1 ).contiguous() , torch.tensor(SCREAMING_SNAKE_CASE ) , ) def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' # Initialise PyTorch model __UpperCAmelCase = ReformerConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) __UpperCAmelCase = ReformerModelWithLMHead(SCREAMING_SNAKE_CASE ) with open(SCREAMING_SNAKE_CASE , '''rb''' ) as f: __UpperCAmelCase = pickle.load(SCREAMING_SNAKE_CASE )['''weights'''] set_model_weights_in_torch(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , config.hidden_size ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": A_ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--trax_model_pkl_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained Reformer model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A_ : List[Any] = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
333
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: while a != 0: snake_case_ , snake_case_ = b % a, a return b def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: if gcd(UpperCAmelCase , UpperCAmelCase ) != 1: snake_case_ = f'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = 1, 0, a snake_case_ , snake_case_ , snake_case_ = 0, 1, m while va != 0: snake_case_ = ua // va snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
69
0
import sys from collections import defaultdict class a : """simple docstring""" def __init__( self : Optional[int] ) -> Optional[int]: __snake_case : List[str] = [] def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[str] ) -> Any: return self.node_position[vertex] def __snake_case ( self : str , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Dict: __snake_case : List[str] = pos def __snake_case ( self : Optional[int] , lowerCamelCase : Union[str, Any] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Dict ) -> str: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: __snake_case : int = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: __snake_case : List[str] = 2 * start + 1 else: __snake_case : Dict = 2 * start + 2 if heap[smallest_child] < heap[start]: __snake_case , __snake_case : Union[str, Any] = heap[smallest_child], positions[smallest_child] __snake_case , __snake_case : str = ( heap[start], positions[start], ) __snake_case , __snake_case : Union[str, Any] = temp, tempa __snake_case : Optional[int] = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] , self.get_position(positions[start] ) ) self.set_position(positions[start] , lowerCAmelCase__ ) self.top_to_bottom(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __snake_case ( self : Dict , lowerCamelCase : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[str]: __snake_case : List[str] = position[index] while index != 0: __snake_case : List[Any] = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: __snake_case : List[Any] = heap[parent] __snake_case : Optional[int] = position[parent] self.set_position(position[parent] , lowerCAmelCase__ ) else: __snake_case : List[Any] = val __snake_case : str = temp self.set_position(lowerCAmelCase__ , lowerCAmelCase__ ) break __snake_case : Union[str, Any] = parent else: __snake_case : Dict = val __snake_case : Union[str, Any] = temp self.set_position(lowerCAmelCase__ , 0 ) def __snake_case ( self : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Any ) -> Dict: __snake_case : Tuple = len(lowerCAmelCase__ ) // 2 - 1 for i in range(lowerCAmelCase__ , -1 , -1 ): self.top_to_bottom(lowerCAmelCase__ , lowerCAmelCase__ , len(lowerCAmelCase__ ) , lowerCAmelCase__ ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> Union[str, Any]: __snake_case : Dict = positions[0] __snake_case : List[Any] = sys.maxsize self.top_to_bottom(lowerCAmelCase__ , 0 , len(lowerCAmelCase__ ) , lowerCAmelCase__ ) return temp def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Dict = Heap() __snake_case : Tuple = [0] * len(__lowerCamelCase ) __snake_case : Optional[int] = [-1] * len(__lowerCamelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph __snake_case : int = [] # Heap of Distance of vertices from their neighboring vertex __snake_case : List[str] = [] for vertex in range(len(__lowerCamelCase ) ): distance_tv.append(sys.maxsize ) positions.append(__lowerCamelCase ) heap.node_position.append(__lowerCamelCase ) __snake_case : List[Any] = [] __snake_case : Optional[int] = 1 __snake_case : List[Any] = sys.maxsize for neighbor, distance in adjacency_list[0]: __snake_case : List[str] = 0 __snake_case : Any = distance heap.heapify(__lowerCamelCase , __lowerCamelCase ) for _ in range(1 , len(__lowerCamelCase ) ): __snake_case : Optional[int] = heap.delete_minimum(__lowerCamelCase , __lowerCamelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) __snake_case : Union[str, Any] = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__lowerCamelCase )] ): __snake_case : Tuple = distance heap.bottom_to_top( __lowerCamelCase , heap.get_position(__lowerCamelCase ) , __lowerCamelCase , __lowerCamelCase ) __snake_case : Any = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _snake_case : int = int(input("Enter number of edges: ").strip()) _snake_case : Optional[int] = defaultdict(list) for _ in range(edges_number): _snake_case : str = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
123
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase = { '''configuration_bert''': ['''BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BertConfig''', '''BertOnnxConfig'''], '''tokenization_bert''': ['''BasicTokenizer''', '''BertTokenizer''', '''WordpieceTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''BertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BertForMaskedLM''', '''BertForMultipleChoice''', '''BertForNextSentencePrediction''', '''BertForPreTraining''', '''BertForQuestionAnswering''', '''BertForSequenceClassification''', '''BertForTokenClassification''', '''BertLayer''', '''BertLMHeadModel''', '''BertModel''', '''BertPreTrainedModel''', '''load_tf_weights_in_bert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBertEmbeddings''', '''TFBertForMaskedLM''', '''TFBertForMultipleChoice''', '''TFBertForNextSentencePrediction''', '''TFBertForPreTraining''', '''TFBertForQuestionAnswering''', '''TFBertForSequenceClassification''', '''TFBertForTokenClassification''', '''TFBertLMHeadModel''', '''TFBertMainLayer''', '''TFBertModel''', '''TFBertPreTrainedModel''', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFBertTokenizer'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxBertForCausalLM''', '''FlaxBertForMaskedLM''', '''FlaxBertForMultipleChoice''', '''FlaxBertForNextSentencePrediction''', '''FlaxBertForPreTraining''', '''FlaxBertForQuestionAnswering''', '''FlaxBertForSequenceClassification''', '''FlaxBertForTokenClassification''', '''FlaxBertModel''', '''FlaxBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
from pathlib import Path import fire from tqdm import tqdm def A ( _SCREAMING_SNAKE_CASE="ro" ,_SCREAMING_SNAKE_CASE="en" ,_SCREAMING_SNAKE_CASE="wmt16" ,_SCREAMING_SNAKE_CASE=None ) -> None: try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("run pip install datasets" ) lowerCamelCase : Union[str, Any] = f'''{src_lang}-{tgt_lang}''' print(f'''Converting {dataset}-{pair}''' ) lowerCamelCase : str = datasets.load_dataset(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) if save_dir is None: lowerCamelCase : Dict = f'''{dataset}-{pair}''' lowerCamelCase : int = Path(_SCREAMING_SNAKE_CASE ) save_dir.mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) for split in ds.keys(): print(f'''Splitting {split} with {ds[split].num_rows} records''' ) # to save to val.source, val.target like summary datasets lowerCamelCase : Any = "val" if split == "validation" else split lowerCamelCase : Dict = save_dir.joinpath(f'''{fn}.source''' ) lowerCamelCase : int = save_dir.joinpath(f'''{fn}.target''' ) lowerCamelCase : int = src_path.open("w+" ) lowerCamelCase : str = tgt_path.open("w+" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): lowerCamelCase : Union[str, Any] = x["translation"] src_fp.write(ex[src_lang] + "\n" ) tgt_fp.write(ex[tgt_lang] + "\n" ) print(f'''Saved {dataset} dataset to {save_dir}''' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
48
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.linear_k''': '''encoder.layers.*.self_attn.linear_k''', '''self_attn.linear_v''': '''encoder.layers.*.self_attn.linear_v''', '''self_attn.linear_q''': '''encoder.layers.*.self_attn.linear_q''', '''self_attn.pos_bias_u''': '''encoder.layers.*.self_attn.pos_bias_u''', '''self_attn.pos_bias_v''': '''encoder.layers.*.self_attn.pos_bias_v''', '''self_attn.linear_out''': '''encoder.layers.*.self_attn.linear_out''', '''self_attn.linear_pos''': '''encoder.layers.*.self_attn.linear_pos''', '''self_attn.rotary_emb''': '''encoder.embed_positions''', '''self_attn_layer_norm''': '''encoder.layers.*.self_attn_layer_norm''', '''conv_module.pointwise_conv1''': '''encoder.layers.*.conv_module.pointwise_conv1''', '''conv_module.pointwise_conv2''': '''encoder.layers.*.conv_module.pointwise_conv2''', '''conv_module.depthwise_conv''': '''encoder.layers.*.conv_module.depthwise_conv''', '''conv_module.batch_norm''': '''encoder.layers.*.conv_module.batch_norm''', '''conv_module.layer_norm''': '''encoder.layers.*.conv_module.layer_norm''', '''ffn1.w_1''': '''encoder.layers.*.ffn1.intermediate_dense''', '''ffn1.w_2''': '''encoder.layers.*.ffn1.output_dense''', '''ffn1.layer_norm''': '''encoder.layers.*.ffn1_layer_norm''', '''ffn2.w_1''': '''encoder.layers.*.ffn2.intermediate_dense''', '''ffn2.w_2''': '''encoder.layers.*.ffn2.output_dense''', '''ffn2.layer_norm''': '''encoder.layers.*.ffn2_layer_norm''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __UpperCamelCase = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: for attribute in key.split('.' ): snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ) if weight_type is not None: snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ).shape else: snake_case_ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": snake_case_ = value elif weight_type == "weight_g": snake_case_ = value elif weight_type == "weight_v": snake_case_ = value elif weight_type == "bias": snake_case_ = value elif weight_type == "running_mean": snake_case_ = value elif weight_type == "running_var": snake_case_ = value elif weight_type == "num_batches_tracked": snake_case_ = value elif weight_type == "inv_freq": snake_case_ = value else: snake_case_ = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = [] snake_case_ = fairseq_model.state_dict() snake_case_ = hf_model.wavaveca_conformer.feature_extractor for name, value in fairseq_dict.items(): snake_case_ = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) snake_case_ = True else: for key, mapped_key in MAPPING.items(): snake_case_ = 'wav2vec2_conformer.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case_ = True if "*" in mapped_key: snake_case_ = name.split(UpperCAmelCase )[0].split('.' )[-2] snake_case_ = mapped_key.replace('*' , UpperCAmelCase ) if "pos_bias_u" in name: snake_case_ = None elif "pos_bias_v" in name: snake_case_ = None elif "weight_g" in name: snake_case_ = 'weight_g' elif "weight_v" in name: snake_case_ = 'weight_v' elif "bias" in name: snake_case_ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case_ = 'weight' elif "running_mean" in name: snake_case_ = 'running_mean' elif "inv_freq" in name: snake_case_ = 'inv_freq' elif "running_var" in name: snake_case_ = 'running_var' elif "num_batches_tracked" in name: snake_case_ = 'num_batches_tracked' else: snake_case_ = None set_recursively(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) continue if not is_used: unused_weights.append(UpperCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: snake_case_ = full_name.split('conv_layers.' )[-1] snake_case_ = name.split('.' ) snake_case_ = int(items[0] ) snake_case_ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase ) @torch.no_grad() def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=True ) -> str: if config_path is not None: snake_case_ = WavaVecaConformerConfig.from_pretrained(UpperCAmelCase , hidden_act='swish' ) else: snake_case_ = WavaVecaConformerConfig() if "rope" in checkpoint_path: snake_case_ = 'rotary' if is_finetuned: if dict_path: snake_case_ = Dictionary.load(UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq snake_case_ = target_dict.pad_index snake_case_ = target_dict.bos_index snake_case_ = target_dict.eos_index snake_case_ = len(target_dict.symbols ) snake_case_ = os.path.join(UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase ) ) return os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) snake_case_ = target_dict.indices # fairseq has the <pad> and <s> switched snake_case_ = 0 snake_case_ = 1 with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(UpperCAmelCase , UpperCAmelCase ) snake_case_ = WavaVecaCTCTokenizer( UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase , ) snake_case_ = True if config.feat_extract_norm == 'layer' else False snake_case_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=UpperCAmelCase , return_attention_mask=UpperCAmelCase , ) snake_case_ = WavaVecaProcessor(feature_extractor=UpperCAmelCase , tokenizer=UpperCAmelCase ) processor.save_pretrained(UpperCAmelCase ) snake_case_ = WavaVecaConformerForCTC(UpperCAmelCase ) else: snake_case_ = WavaVecaConformerForPreTraining(UpperCAmelCase ) if is_finetuned: snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: snake_case_ = argparse.Namespace(task='audio_pretraining' ) snake_case_ = fairseq.tasks.setup_task(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=UpperCAmelCase ) snake_case_ = model[0].eval() recursively_load_weights(UpperCAmelCase , UpperCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __UpperCamelCase = parser.parse_args() convert_wavaveca_conformer_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
69
0
import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import OpenLlamaForCausalLM, OpenLlamaForSequenceClassification, OpenLlamaModel class __lowercase : """simple docstring""" def __init__( self : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[Any]=13 , lowerCAmelCase__ : Union[str, Any]=7 , lowerCAmelCase__ : Union[str, Any]=True , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : Any=False , lowerCAmelCase__ : int=True , lowerCAmelCase__ : List[Any]=99 , lowerCAmelCase__ : str=32 , lowerCAmelCase__ : Union[str, Any]=5 , lowerCAmelCase__ : int=4 , lowerCAmelCase__ : Any=37 , lowerCAmelCase__ : Dict="gelu" , lowerCAmelCase__ : str=0.1 , lowerCAmelCase__ : Optional[Any]=0.1 , lowerCAmelCase__ : Dict=512 , lowerCAmelCase__ : Optional[int]=16 , lowerCAmelCase__ : List[str]=2 , lowerCAmelCase__ : Optional[int]=0.02 , lowerCAmelCase__ : Any=3 , lowerCAmelCase__ : str=4 , lowerCAmelCase__ : Dict=None , ): SCREAMING_SNAKE_CASE_: Union[str, Any] = parent SCREAMING_SNAKE_CASE_: Tuple = batch_size SCREAMING_SNAKE_CASE_: Union[str, Any] = seq_length SCREAMING_SNAKE_CASE_: Optional[int] = is_training SCREAMING_SNAKE_CASE_: Dict = use_input_mask SCREAMING_SNAKE_CASE_: Optional[int] = use_token_type_ids SCREAMING_SNAKE_CASE_: Optional[int] = use_labels SCREAMING_SNAKE_CASE_: List[str] = vocab_size SCREAMING_SNAKE_CASE_: Optional[int] = hidden_size SCREAMING_SNAKE_CASE_: Dict = num_hidden_layers SCREAMING_SNAKE_CASE_: Any = num_attention_heads SCREAMING_SNAKE_CASE_: Union[str, Any] = intermediate_size SCREAMING_SNAKE_CASE_: str = hidden_act SCREAMING_SNAKE_CASE_: Tuple = hidden_dropout_prob SCREAMING_SNAKE_CASE_: Tuple = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_: Any = max_position_embeddings SCREAMING_SNAKE_CASE_: Tuple = type_vocab_size SCREAMING_SNAKE_CASE_: Union[str, Any] = type_sequence_label_size SCREAMING_SNAKE_CASE_: Any = initializer_range SCREAMING_SNAKE_CASE_: Any = num_labels SCREAMING_SNAKE_CASE_: Union[str, Any] = num_choices SCREAMING_SNAKE_CASE_: List[Any] = scope def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_: str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) SCREAMING_SNAKE_CASE_: List[Any] = None if self.use_input_mask: SCREAMING_SNAKE_CASE_: Optional[int] = random_attention_mask([self.batch_size, self.seq_length]) SCREAMING_SNAKE_CASE_: str = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE_: List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) SCREAMING_SNAKE_CASE_: Optional[Any] = None SCREAMING_SNAKE_CASE_: Optional[int] = None SCREAMING_SNAKE_CASE_: Union[str, Any] = None if self.use_labels: SCREAMING_SNAKE_CASE_: Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size) SCREAMING_SNAKE_CASE_: List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) SCREAMING_SNAKE_CASE_: int = ids_tensor([self.batch_size] , self.num_choices) SCREAMING_SNAKE_CASE_: str = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self : Any): return OpenLlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , use_stable_embedding=lowerCAmelCase__ , ) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase__ : str , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Any , lowerCAmelCase__ : int , lowerCAmelCase__ : int): SCREAMING_SNAKE_CASE_: List[str] = OpenLlamaModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Optional[int] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = model(lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : int , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : List[Any] , ): SCREAMING_SNAKE_CASE_: List[Any] = True SCREAMING_SNAKE_CASE_: Optional[int] = OpenLlamaModel(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Tuple = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_: Optional[int] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_: Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : str , lowerCAmelCase__ : int , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : int , lowerCAmelCase__ : str , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : int , ): SCREAMING_SNAKE_CASE_: Optional[Any] = OpenLlamaForCausalLM(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Union[str, Any] , ): SCREAMING_SNAKE_CASE_: str = True SCREAMING_SNAKE_CASE_: Optional[Any] = True SCREAMING_SNAKE_CASE_: Optional[Any] = OpenLlamaForCausalLM(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() # first forward pass SCREAMING_SNAKE_CASE_: Optional[int] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_: Any = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids SCREAMING_SNAKE_CASE_: List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size) SCREAMING_SNAKE_CASE_: List[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2) # append to next input_ids and SCREAMING_SNAKE_CASE_: str = torch.cat([input_ids, next_tokens] , dim=-1) SCREAMING_SNAKE_CASE_: Any = torch.cat([input_mask, next_mask] , dim=-1) SCREAMING_SNAKE_CASE_: Tuple = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["hidden_states"][0] SCREAMING_SNAKE_CASE_: List[str] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["hidden_states"][0] # select random slice SCREAMING_SNAKE_CASE_: Union[str, Any] = ids_tensor((1,) , output_from_past.shape[-1]).item() SCREAMING_SNAKE_CASE_: Tuple = output_from_no_past[:, -3:, random_slice_idx].detach() SCREAMING_SNAKE_CASE_: List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3)) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: Optional[int] = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ): str = config_and_inputs SCREAMING_SNAKE_CASE_: Any = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __lowercase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Dict = ( (OpenLlamaModel, OpenLlamaForCausalLM, OpenLlamaForSequenceClassification) if is_torch_available() else () ) _UpperCAmelCase : int = (OpenLlamaForCausalLM,) if is_torch_available() else () _UpperCAmelCase : Union[str, Any] = ( { '''feature-extraction''': OpenLlamaModel, '''text-classification''': OpenLlamaForSequenceClassification, '''text-generation''': OpenLlamaForCausalLM, '''zero-shot''': OpenLlamaForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase : List[Any] = False _UpperCAmelCase : Union[str, Any] = False def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_: Optional[Any] = OpenLlamaModelTester(self) SCREAMING_SNAKE_CASE_: Dict = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=37) def _SCREAMING_SNAKE_CASE ( self : List[str]): self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: Optional[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: SCREAMING_SNAKE_CASE_: Optional[Any] = type self.model_tester.create_and_check_model(*lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_: Any = 3 SCREAMING_SNAKE_CASE_: Any = input_dict["input_ids"] SCREAMING_SNAKE_CASE_: List[Any] = input_ids.ne(1).to(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) SCREAMING_SNAKE_CASE_: Any = OpenLlamaForSequenceClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: str = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_: Any = 3 SCREAMING_SNAKE_CASE_: str = "single_label_classification" SCREAMING_SNAKE_CASE_: Any = input_dict["input_ids"] SCREAMING_SNAKE_CASE_: Optional[Any] = input_ids.ne(1).to(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) SCREAMING_SNAKE_CASE_: int = OpenLlamaForSequenceClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_: Dict = 3 SCREAMING_SNAKE_CASE_: Optional[int] = "multi_label_classification" SCREAMING_SNAKE_CASE_: Dict = input_dict["input_ids"] SCREAMING_SNAKE_CASE_: str = input_ids.ne(1).to(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) SCREAMING_SNAKE_CASE_: Any = OpenLlamaForSequenceClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: List[Any] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @unittest.skip("Open-Llama buffers include complex numbers, which breaks this test") def _SCREAMING_SNAKE_CASE ( self : Any): pass @parameterized.expand([("linear",), ("dynamic",)]) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase__ : List[str]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_: str = ids_tensor([1, 10] , config.vocab_size) SCREAMING_SNAKE_CASE_: Tuple = ids_tensor([1, int(config.max_position_embeddings * 1.5)] , config.vocab_size) set_seed(42) # Fixed seed at init time so the two models get the same random weights SCREAMING_SNAKE_CASE_: Optional[int] = OpenLlamaModel(lowerCAmelCase__) original_model.to(lowerCAmelCase__) original_model.eval() SCREAMING_SNAKE_CASE_: int = original_model(lowerCAmelCase__).last_hidden_state SCREAMING_SNAKE_CASE_: str = original_model(lowerCAmelCase__).last_hidden_state set_seed(42) # Fixed seed at init time so the two models get the same random weights SCREAMING_SNAKE_CASE_: Optional[int] = {"type": scaling_type, "factor": 10.0} SCREAMING_SNAKE_CASE_: Dict = OpenLlamaModel(lowerCAmelCase__) scaled_model.to(lowerCAmelCase__) scaled_model.eval() SCREAMING_SNAKE_CASE_: List[Any] = scaled_model(lowerCAmelCase__).last_hidden_state SCREAMING_SNAKE_CASE_: Optional[int] = scaled_model(lowerCAmelCase__).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5)) else: self.assertFalse(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5)) # The output should be different for long inputs self.assertFalse(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5))
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase ) -> list: if len(UpperCAmelCase ) <= 1: return [tuple(UpperCAmelCase )] snake_case_ = [] def generate(UpperCAmelCase , UpperCAmelCase ): snake_case_ = [0] * n res.append(tuple(UpperCAmelCase ) ) snake_case_ = 0 while i < n: if c[i] < i: if i % 2 == 0: snake_case_ , snake_case_ = arr[i], arr[0] else: snake_case_ , snake_case_ = arr[i], arr[c[i]] res.append(tuple(UpperCAmelCase ) ) c[i] += 1 snake_case_ = 0 else: snake_case_ = 0 i += 1 generate(len(UpperCAmelCase ) , UpperCAmelCase ) return res if __name__ == "__main__": __UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
69
0
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency lowerCamelCase__ = { """E""": 12.70, """T""": 9.06, """A""": 8.17, """O""": 7.51, """I""": 6.97, """N""": 6.75, """S""": 6.33, """H""": 6.09, """R""": 5.99, """D""": 4.25, """L""": 4.03, """C""": 2.78, """U""": 2.76, """M""": 2.41, """W""": 2.36, """F""": 2.23, """G""": 2.02, """Y""": 1.97, """P""": 1.93, """B""": 1.29, """V""": 0.98, """K""": 0.77, """J""": 0.15, """X""": 0.15, """Q""": 0.10, """Z""": 0.07, } lowerCamelCase__ = """ETAOINSHRDLCUMWFGYPBVKJXQZ""" lowerCamelCase__ = """ABCDEFGHIJKLMNOPQRSTUVWXYZ""" def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : int ): """simple docstring""" __a = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" return x[0] def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" __a = get_letter_count(_SCREAMING_SNAKE_CASE ) __a = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(_SCREAMING_SNAKE_CASE ) __a = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=_SCREAMING_SNAKE_CASE ) __a = """""".join(freq_to_letter[freq] ) __a = list(freq_to_letter_str.items() ) freq_pairs.sort(key=_SCREAMING_SNAKE_CASE , reverse=_SCREAMING_SNAKE_CASE ) __a = [freq_pair[1] for freq_pair in freq_pairs] return "".join(_SCREAMING_SNAKE_CASE ) def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Tuple ): """simple docstring""" __a = get_frequency_order(_SCREAMING_SNAKE_CASE ) __a = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __UpperCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __UpperCamelCase = TaTokenizerFast __UpperCamelCase = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __UpperCamelCase = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
69
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging A__: Union[str, Any] = logging.get_logger(__name__) A__: Union[str, Any] = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class A__ ( lowerCAmelCase__ ): __UpperCamelCase : Dict = "gpt_bigcode" __UpperCamelCase : int = ["past_key_values"] __UpperCamelCase : Union[str, Any] = { "hidden_size": "n_embd", "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self :Tuple , SCREAMING_SNAKE_CASE :Tuple=5_0_2_5_7 , SCREAMING_SNAKE_CASE :Dict=1_0_2_4 , SCREAMING_SNAKE_CASE :Any=7_6_8 , SCREAMING_SNAKE_CASE :Union[str, Any]=1_2 , SCREAMING_SNAKE_CASE :List[str]=1_2 , SCREAMING_SNAKE_CASE :Tuple=None , SCREAMING_SNAKE_CASE :Dict="gelu_pytorch_tanh" , SCREAMING_SNAKE_CASE :Dict=0.1 , SCREAMING_SNAKE_CASE :Any=0.1 , SCREAMING_SNAKE_CASE :Optional[Any]=0.1 , SCREAMING_SNAKE_CASE :List[Any]=1e-5 , SCREAMING_SNAKE_CASE :List[Any]=0.02 , SCREAMING_SNAKE_CASE :Optional[int]=True , SCREAMING_SNAKE_CASE :Optional[Any]=True , SCREAMING_SNAKE_CASE :Union[str, Any]=5_0_2_5_6 , SCREAMING_SNAKE_CASE :List[str]=5_0_2_5_6 , SCREAMING_SNAKE_CASE :Optional[Any]=True , SCREAMING_SNAKE_CASE :List[str]=True , SCREAMING_SNAKE_CASE :List[str]=True , **SCREAMING_SNAKE_CASE :Optional[int] , ) -> Any: '''simple docstring''' _a : Any =vocab_size _a : int =n_positions _a : Any =n_embd _a : Any =n_layer _a : Tuple =n_head _a : List[Any] =n_inner _a : Optional[int] =activation_function _a : Optional[Any] =resid_pdrop _a : List[str] =embd_pdrop _a : int =attn_pdrop _a : Union[str, Any] =layer_norm_epsilon _a : List[Any] =initializer_range _a : Union[str, Any] =scale_attn_weights _a : List[Any] =use_cache _a : Any =attention_softmax_in_fpaa _a : List[str] =scale_attention_softmax_in_fpaa _a : Optional[Any] =multi_query _a : Optional[int] =bos_token_id _a : Any =eos_token_id super().__init__(bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ )
276
"""simple docstring""" import sys from collections import defaultdict class UpperCamelCase : def __init__( self) -> Optional[int]: snake_case_ = [] def a_ ( self, lowerCAmelCase__) -> Any: return self.node_position[vertex] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = pos def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> str: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: snake_case_ = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: snake_case_ = 2 * start + 1 else: snake_case_ = 2 * start + 2 if heap[smallest_child] < heap[start]: snake_case_ , snake_case_ = heap[smallest_child], positions[smallest_child] snake_case_ , snake_case_ = ( heap[start], positions[start], ) snake_case_ , snake_case_ = temp, tempa snake_case_ = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child], self.get_position(positions[start])) self.set_position(positions[start], lowerCAmelCase__) self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> List[str]: snake_case_ = position[index] while index != 0: snake_case_ = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: snake_case_ = heap[parent] snake_case_ = position[parent] self.set_position(position[parent], lowerCAmelCase__) else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, lowerCAmelCase__) break snake_case_ = parent else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, 0) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = len(lowerCAmelCase__) // 2 - 1 for i in range(lowerCAmelCase__, -1, -1): self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, len(lowerCAmelCase__), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = positions[0] snake_case_ = sys.maxsize self.top_to_bottom(lowerCAmelCase__, 0, len(lowerCAmelCase__), lowerCAmelCase__) return temp def UpperCAmelCase ( UpperCAmelCase ) -> Tuple: snake_case_ = Heap() snake_case_ = [0] * len(UpperCAmelCase ) snake_case_ = [-1] * len(UpperCAmelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph snake_case_ = [] # Heap of Distance of vertices from their neighboring vertex snake_case_ = [] for vertex in range(len(UpperCAmelCase ) ): distance_tv.append(sys.maxsize ) positions.append(UpperCAmelCase ) heap.node_position.append(UpperCAmelCase ) snake_case_ = [] snake_case_ = 1 snake_case_ = sys.maxsize for neighbor, distance in adjacency_list[0]: snake_case_ = 0 snake_case_ = distance heap.heapify(UpperCAmelCase , UpperCAmelCase ) for _ in range(1 , len(UpperCAmelCase ) ): snake_case_ = heap.delete_minimum(UpperCAmelCase , UpperCAmelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) snake_case_ = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(UpperCAmelCase )] ): snake_case_ = distance heap.bottom_to_top( UpperCAmelCase , heap.get_position(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase ) snake_case_ = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > __UpperCamelCase = int(input('''Enter number of edges: ''').strip()) __UpperCamelCase = defaultdict(list) for _ in range(edges_number): __UpperCamelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
69
0
'''simple docstring''' from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class a__ : pass
250
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase = { '''configuration_mctct''': ['''MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MCTCTConfig'''], '''feature_extraction_mctct''': ['''MCTCTFeatureExtractor'''], '''processing_mctct''': ['''MCTCTProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MCTCTForCTC''', '''MCTCTModel''', '''MCTCTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
'''simple docstring''' def __UpperCAmelCase ( a_: Union[str, Any], a_: Dict ): if not (isinstance(a_, a_ ) and isinstance(a_, a_ )): raise ValueError("longest_common_substring() takes two strings for inputs" ) _UpperCAmelCase : List[str] = len(a_ ) _UpperCAmelCase : List[Any] = len(a_ ) _UpperCAmelCase : Dict = [[0] * (texta_length + 1) for _ in range(texta_length + 1 )] _UpperCAmelCase : str = 0 _UpperCAmelCase : Optional[Any] = 0 for i in range(1, texta_length + 1 ): for j in range(1, texta_length + 1 ): if texta[i - 1] == texta[j - 1]: _UpperCAmelCase : Dict = 1 + dp[i - 1][j - 1] if dp[i][j] > ans_length: _UpperCAmelCase : Tuple = i _UpperCAmelCase : Optional[int] = dp[i][j] return texta[ans_index - ans_length : ans_index] if __name__ == "__main__": import doctest doctest.testmod()
145
"""simple docstring""" from math import factorial def UpperCAmelCase ( UpperCAmelCase = 20 ) -> int: snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(UpperCAmelCase ) / (factorial(UpperCAmelCase ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: __UpperCamelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
69
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase_ = { """configuration_squeezebert""": [ """SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SqueezeBertConfig""", """SqueezeBertOnnxConfig""", ], """tokenization_squeezebert""": ["""SqueezeBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""SqueezeBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """SqueezeBertForMaskedLM""", """SqueezeBertForMultipleChoice""", """SqueezeBertForQuestionAnswering""", """SqueezeBertForSequenceClassification""", """SqueezeBertForTokenClassification""", """SqueezeBertModel""", """SqueezeBertModule""", """SqueezeBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
303
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
0
"""simple docstring""" import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline _A : Dict = version.parse(version.parse(torch.__version__).base_version) < version.parse("""1.11""") def __magic_name__ ( __snake_case : str , __snake_case : List[Any] , __snake_case : int , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : Union[str, Any] , __snake_case : Dict=False , ) -> Optional[Any]: output_path.parent.mkdir(parents=__snake_case , exist_ok=__snake_case ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( __snake_case , __snake_case , f=output_path.as_posix() , input_names=__snake_case , output_names=__snake_case , dynamic_axes=__snake_case , do_constant_folding=__snake_case , use_external_data_format=__snake_case , enable_onnx_checker=__snake_case , opset_version=__snake_case , ) else: export( __snake_case , __snake_case , f=output_path.as_posix() , input_names=__snake_case , output_names=__snake_case , dynamic_axes=__snake_case , do_constant_folding=__snake_case , opset_version=__snake_case , ) @torch.no_grad() def __magic_name__ ( __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[str] , __snake_case : List[str] = False ) -> Tuple: lowercase : Dict = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): lowercase : Optional[int] = "cuda" elif fpaa and not torch.cuda.is_available(): raise ValueError("`float16` model export is only supported on GPUs with CUDA" ) else: lowercase : Optional[Any] = "cpu" lowercase : Optional[Any] = StableDiffusionPipeline.from_pretrained(__snake_case , torch_dtype=__snake_case ).to(__snake_case ) lowercase : Optional[int] = Path(__snake_case ) # TEXT ENCODER lowercase : Dict = pipeline.text_encoder.config.max_position_embeddings lowercase : Optional[Any] = pipeline.text_encoder.config.hidden_size lowercase : Dict = pipeline.tokenizer( "A sample prompt" , padding="max_length" , max_length=pipeline.tokenizer.model_max_length , truncation=__snake_case , return_tensors="pt" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=__snake_case , dtype=torch.intaa )) , output_path=output_path / "text_encoder" / "model.onnx" , ordered_input_names=["input_ids"] , output_names=["last_hidden_state", "pooler_output"] , dynamic_axes={ "input_ids": {0: "batch", 1: "sequence"}, } , opset=__snake_case , ) del pipeline.text_encoder # UNET lowercase : Tuple = pipeline.unet.config.in_channels lowercase : Optional[int] = pipeline.unet.config.sample_size lowercase : str = output_path / "unet" / "model.onnx" onnx_export( pipeline.unet , model_args=( torch.randn(2 , __snake_case , __snake_case , __snake_case ).to(device=__snake_case , dtype=__snake_case ), torch.randn(2 ).to(device=__snake_case , dtype=__snake_case ), torch.randn(2 , __snake_case , __snake_case ).to(device=__snake_case , dtype=__snake_case ), False, ) , output_path=__snake_case , ordered_input_names=["sample", "timestep", "encoder_hidden_states", "return_dict"] , output_names=["out_sample"] , dynamic_axes={ "sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, "timestep": {0: "batch"}, "encoder_hidden_states": {0: "batch", 1: "sequence"}, } , opset=__snake_case , use_external_data_format=__snake_case , ) lowercase : int = str(unet_path.absolute().as_posix() ) lowercase : List[str] = os.path.dirname(__snake_case ) lowercase : List[str] = onnx.load(__snake_case ) # clean up existing tensor files shutil.rmtree(__snake_case ) os.mkdir(__snake_case ) # collate external tensor files into one onnx.save_model( __snake_case , __snake_case , save_as_external_data=__snake_case , all_tensors_to_one_file=__snake_case , location="weights.pb" , convert_attribute=__snake_case , ) del pipeline.unet # VAE ENCODER lowercase : Dict = pipeline.vae lowercase : Optional[Any] = vae_encoder.config.in_channels lowercase : Dict = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder lowercase : str = lambda __snake_case , __snake_case : vae_encoder.encode(__snake_case , __snake_case )[0].sample() onnx_export( __snake_case , model_args=( torch.randn(1 , __snake_case , __snake_case , __snake_case ).to(device=__snake_case , dtype=__snake_case ), False, ) , output_path=output_path / "vae_encoder" / "model.onnx" , ordered_input_names=["sample", "return_dict"] , output_names=["latent_sample"] , dynamic_axes={ "sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, } , opset=__snake_case , ) # VAE DECODER lowercase : Dict = pipeline.vae lowercase : Optional[int] = vae_decoder.config.latent_channels lowercase : Union[str, Any] = vae_decoder.config.out_channels # forward only through the decoder part lowercase : Optional[int] = vae_encoder.decode onnx_export( __snake_case , model_args=( torch.randn(1 , __snake_case , __snake_case , __snake_case ).to(device=__snake_case , dtype=__snake_case ), False, ) , output_path=output_path / "vae_decoder" / "model.onnx" , ordered_input_names=["latent_sample", "return_dict"] , output_names=["sample"] , dynamic_axes={ "latent_sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, } , opset=__snake_case , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: lowercase : List[str] = pipeline.safety_checker lowercase : str = safety_checker.config.vision_config.num_channels lowercase : List[Any] = safety_checker.config.vision_config.image_size lowercase : Any = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , __snake_case , __snake_case , __snake_case , ).to(device=__snake_case , dtype=__snake_case ), torch.randn(1 , __snake_case , __snake_case , __snake_case ).to(device=__snake_case , dtype=__snake_case ), ) , output_path=output_path / "safety_checker" / "model.onnx" , ordered_input_names=["clip_input", "images"] , output_names=["out_images", "has_nsfw_concepts"] , dynamic_axes={ "clip_input": {0: "batch", 1: "channels", 2: "height", 3: "width"}, "images": {0: "batch", 1: "height", 2: "width", 3: "channels"}, } , opset=__snake_case , ) del pipeline.safety_checker lowercase : Optional[Any] = OnnxRuntimeModel.from_pretrained(output_path / "safety_checker" ) lowercase : Optional[int] = pipeline.feature_extractor else: lowercase : Union[str, Any] = None lowercase : Dict = None lowercase : List[str] = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / "vae_encoder" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / "vae_decoder" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / "text_encoder" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / "unet" ) , scheduler=pipeline.scheduler , safety_checker=__snake_case , feature_extractor=__snake_case , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(__snake_case ) print("ONNX pipeline saved to" , __snake_case ) del pipeline del onnx_pipeline lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(__snake_case , provider="CPUExecutionProvider" ) print("ONNX pipeline is loadable" ) if __name__ == "__main__": _A : str = argparse.ArgumentParser() parser.add_argument( """--model_path""", type=str, required=True, help="""Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).""", ) parser.add_argument("""--output_path""", type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--opset""", default=14, type=int, help="""The version of the ONNX operator set to use.""", ) parser.add_argument("""--fp16""", action="""store_true""", default=False, help="""Export the models in `float16` mode""") _A : Tuple = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
202
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = (DPMSolverSinglestepScheduler,) SCREAMING_SNAKE_CASE_ = (("num_inference_steps", 2_5),) def a_ ( self, **lowerCAmelCase__) -> int: snake_case_ = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf'), 'variance_type': None, } config.update(**lowerCAmelCase__) return config def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> List[Any]: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ , snake_case_ = sample, sample for t in range(lowerCAmelCase__, time_step + scheduler.config.solver_order + 1): snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self) -> Union[str, Any]: pass def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> int: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self, lowerCAmelCase__=None, **lowerCAmelCase__) -> Union[str, Any]: if scheduler is None: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample return sample def a_ ( self) -> List[Any]: snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = 50 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2574) < 1e-3 def a_ ( self) -> Dict: for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__) def a_ ( self) -> Optional[Any]: # make sure that iterating over schedulers with same config names gives same results # for defaults snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 snake_case_ = DEISMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverMultistepScheduler.from_config(scheduler.config) snake_case_ = UniPCMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverSinglestepScheduler.from_config(scheduler.config) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> str: self.check_over_configs(thresholding=lowerCAmelCase__) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCAmelCase__, prediction_type=lowerCAmelCase__, sample_max_value=lowerCAmelCase__, algorithm_type='dpmsolver++', solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, ) def a_ ( self) -> Tuple: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__) def a_ ( self) -> Optional[int]: for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) snake_case_ = self.full_loop( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) assert not torch.isnan(lowerCAmelCase__).any(), "Samples have nan numbers" def a_ ( self) -> Optional[Any]: self.check_over_configs(lower_order_final=lowerCAmelCase__) self.check_over_configs(lower_order_final=lowerCAmelCase__) def a_ ( self) -> Any: self.check_over_configs(lambda_min_clipped=-float('inf')) self.check_over_configs(lambda_min_clipped=-5.1) def a_ ( self) -> Any: self.check_over_configs(variance_type=lowerCAmelCase__) self.check_over_configs(variance_type='learned_range') def a_ ( self) -> List[Any]: for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=lowerCAmelCase__, time_step=0) def a_ ( self) -> int: snake_case_ = self.full_loop() snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> Dict: snake_case_ = self.full_loop(use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2248) < 1e-3 def a_ ( self) -> Union[str, Any]: snake_case_ = self.full_loop(prediction_type='v_prediction') snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.1453) < 1e-3 def a_ ( self) -> Optional[Any]: snake_case_ = self.full_loop(prediction_type='v_prediction', use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.0649) < 1e-3 def a_ ( self) -> Optional[int]: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(thresholding=lowerCAmelCase__, dynamic_thresholding_ratio=0) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample assert sample.dtype == torch.floataa
69
0
import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def A_ ( A__ , A__=1 ) -> Any: if n_shave_prefix_segments >= 0: return ".".join(path.split('.' )[n_shave_prefix_segments:] ) else: return ".".join(path.split('.' )[:n_shave_prefix_segments] ) def A_ ( A__ , A__=0 ) -> Optional[Any]: a__ : List[str] = [] for old_item in old_list: a__ : str = old_item.replace('in_layers.0' , 'norm1' ) a__ : List[Any] = new_item.replace('in_layers.2' , 'conv1' ) a__ : str = new_item.replace('out_layers.0' , 'norm2' ) a__ : Optional[int] = new_item.replace('out_layers.3' , 'conv2' ) a__ : Optional[int] = new_item.replace('emb_layers.1' , 'time_emb_proj' ) a__ : List[Any] = new_item.replace('skip_connection' , 'conv_shortcut' ) a__ : Tuple = shave_segments(A__ , n_shave_prefix_segments=A__ ) mapping.append({'old': old_item, 'new': new_item} ) return mapping def A_ ( A__ , A__=0 ) -> List[Any]: a__ : Optional[int] = [] for old_item in old_list: a__ : str = old_item a__ : List[Any] = new_item.replace('norm.weight' , 'group_norm.weight' ) a__ : Union[str, Any] = new_item.replace('norm.bias' , 'group_norm.bias' ) a__ : List[Any] = new_item.replace('proj_out.weight' , 'proj_attn.weight' ) a__ : List[Any] = new_item.replace('proj_out.bias' , 'proj_attn.bias' ) a__ : Union[str, Any] = shave_segments(A__ , n_shave_prefix_segments=A__ ) mapping.append({'old': old_item, 'new': new_item} ) return mapping def A_ ( A__ , A__ , A__ , A__=None , A__=None , A__=None ) -> Any: assert isinstance(A__ , A__ ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): a__ : Dict = old_checkpoint[path] a__ : Any = old_tensor.shape[0] // 3 a__ : str = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) a__ : Optional[Any] = old_tensor.shape[0] // config['num_head_channels'] // 3 a__ : int = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) a__ , a__ , a__ : Any = old_tensor.split(channels // num_heads , dim=1 ) a__ : Optional[int] = query.reshape(A__ ) a__ : Union[str, Any] = key.reshape(A__ ) a__ : int = value.reshape(A__ ) for path in paths: a__ : int = path['new'] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here a__ : Dict = new_path.replace('middle_block.0' , 'mid_block.resnets.0' ) a__ : Optional[Any] = new_path.replace('middle_block.1' , 'mid_block.attentions.0' ) a__ : List[str] = new_path.replace('middle_block.2' , 'mid_block.resnets.1' ) if additional_replacements is not None: for replacement in additional_replacements: a__ : str = new_path.replace(replacement['old'] , replacement['new'] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: a__ : List[str] = old_checkpoint[path['old']][:, :, 0] else: a__ : Optional[Any] = old_checkpoint[path['old']] def A_ ( A__ , A__ ) -> Optional[Any]: a__ : Any = {} a__ : Union[str, Any] = checkpoint['time_embed.0.weight'] a__ : Tuple = checkpoint['time_embed.0.bias'] a__ : Optional[Any] = checkpoint['time_embed.2.weight'] a__ : int = checkpoint['time_embed.2.bias'] a__ : List[str] = checkpoint['input_blocks.0.0.weight'] a__ : Tuple = checkpoint['input_blocks.0.0.bias'] a__ : Union[str, Any] = checkpoint['out.0.weight'] a__ : Tuple = checkpoint['out.0.bias'] a__ : int = checkpoint['out.2.weight'] a__ : List[str] = checkpoint['out.2.bias'] # Retrieves the keys for the input blocks only a__ : List[str] = len({'.'.join(layer.split('.' )[:2] ) for layer in checkpoint if 'input_blocks' in layer} ) a__ : Tuple = { layer_id: [key for key in checkpoint if F'input_blocks.{layer_id}' in key] for layer_id in range(A__ ) } # Retrieves the keys for the middle blocks only a__ : Optional[int] = len({'.'.join(layer.split('.' )[:2] ) for layer in checkpoint if 'middle_block' in layer} ) a__ : Optional[int] = { layer_id: [key for key in checkpoint if F'middle_block.{layer_id}' in key] for layer_id in range(A__ ) } # Retrieves the keys for the output blocks only a__ : Any = len({'.'.join(layer.split('.' )[:2] ) for layer in checkpoint if 'output_blocks' in layer} ) a__ : Union[str, Any] = { layer_id: [key for key in checkpoint if F'output_blocks.{layer_id}' in key] for layer_id in range(A__ ) } for i in range(1 , A__ ): a__ : Tuple = (i - 1) // (config['num_res_blocks'] + 1) a__ : Optional[int] = (i - 1) % (config['num_res_blocks'] + 1) a__ : Any = [key for key in input_blocks[i] if F'input_blocks.{i}.0' in key] a__ : Union[str, Any] = [key for key in input_blocks[i] if F'input_blocks.{i}.1' in key] if F'input_blocks.{i}.0.op.weight' in checkpoint: a__ : Union[str, Any] = checkpoint[ F'input_blocks.{i}.0.op.weight' ] a__ : Optional[Any] = checkpoint[ F'input_blocks.{i}.0.op.bias' ] continue a__ : Dict = renew_resnet_paths(A__ ) a__ : Optional[Any] = {'old': F'input_blocks.{i}.0', 'new': F'down_blocks.{block_id}.resnets.{layer_in_block_id}'} a__ : List[str] = {'old': 'resnets.2.op', 'new': 'downsamplers.0.op'} assign_to_checkpoint( A__ , A__ , A__ , additional_replacements=[meta_path, resnet_op] , config=A__ ) if len(A__ ): a__ : Union[str, Any] = renew_attention_paths(A__ ) a__ : str = { 'old': F'input_blocks.{i}.1', 'new': F'down_blocks.{block_id}.attentions.{layer_in_block_id}', } a__ : List[str] = { F'input_blocks.{i}.1.qkv.bias': { 'key': F'down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias', 'query': F'down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias', 'value': F'down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias', }, F'input_blocks.{i}.1.qkv.weight': { 'key': F'down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight', 'query': F'down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight', 'value': F'down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight', }, } assign_to_checkpoint( A__ , A__ , A__ , additional_replacements=[meta_path] , attention_paths_to_split=A__ , config=A__ , ) a__ : List[Any] = middle_blocks[0] a__ : str = middle_blocks[1] a__ : Optional[int] = middle_blocks[2] a__ : Dict = renew_resnet_paths(A__ ) assign_to_checkpoint(A__ , A__ , A__ , config=A__ ) a__ : List[str] = renew_resnet_paths(A__ ) assign_to_checkpoint(A__ , A__ , A__ , config=A__ ) a__ : int = renew_attention_paths(A__ ) a__ : Tuple = { 'middle_block.1.qkv.bias': { 'key': 'mid_block.attentions.0.key.bias', 'query': 'mid_block.attentions.0.query.bias', 'value': 'mid_block.attentions.0.value.bias', }, 'middle_block.1.qkv.weight': { 'key': 'mid_block.attentions.0.key.weight', 'query': 'mid_block.attentions.0.query.weight', 'value': 'mid_block.attentions.0.value.weight', }, } assign_to_checkpoint( A__ , A__ , A__ , attention_paths_to_split=A__ , config=A__ ) for i in range(A__ ): a__ : Optional[int] = i // (config['num_res_blocks'] + 1) a__ : Optional[int] = i % (config['num_res_blocks'] + 1) a__ : Tuple = [shave_segments(A__ , 2 ) for name in output_blocks[i]] a__ : int = {} for layer in output_block_layers: a__ , a__ : Tuple = layer.split('.' )[0], shave_segments(A__ , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(A__ ) else: a__ : Union[str, Any] = [layer_name] if len(A__ ) > 1: a__ : Tuple = [key for key in output_blocks[i] if F'output_blocks.{i}.0' in key] a__ : Union[str, Any] = [key for key in output_blocks[i] if F'output_blocks.{i}.1' in key] a__ : int = renew_resnet_paths(A__ ) a__ : str = renew_resnet_paths(A__ ) a__ : str = {'old': F'output_blocks.{i}.0', 'new': F'up_blocks.{block_id}.resnets.{layer_in_block_id}'} assign_to_checkpoint(A__ , A__ , A__ , additional_replacements=[meta_path] , config=A__ ) if ["conv.weight", "conv.bias"] in output_block_list.values(): a__ : Union[str, Any] = list(output_block_list.values() ).index(['conv.weight', 'conv.bias'] ) a__ : Optional[Any] = checkpoint[ F'output_blocks.{i}.{index}.conv.weight' ] a__ : List[str] = checkpoint[ F'output_blocks.{i}.{index}.conv.bias' ] # Clear attentions as they have been attributed above. if len(A__ ) == 2: a__ : Optional[int] = [] if len(A__ ): a__ : List[str] = renew_attention_paths(A__ ) a__ : List[Any] = { 'old': F'output_blocks.{i}.1', 'new': F'up_blocks.{block_id}.attentions.{layer_in_block_id}', } a__ : Optional[Any] = { F'output_blocks.{i}.1.qkv.bias': { 'key': F'up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias', 'query': F'up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias', 'value': F'up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias', }, F'output_blocks.{i}.1.qkv.weight': { 'key': F'up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight', 'query': F'up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight', 'value': F'up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight', }, } assign_to_checkpoint( A__ , A__ , A__ , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any('qkv' in key for key in attentions ) else None , config=A__ , ) else: a__ : Optional[Any] = renew_resnet_paths(A__ , n_shave_prefix_segments=1 ) for path in resnet_0_paths: a__ : str = '.'.join(['output_blocks', str(A__ ), path['old']] ) a__ : Union[str, Any] = '.'.join(['up_blocks', str(A__ ), 'resnets', str(A__ ), path['new']] ) a__ : List[str] = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") lowercase : str = parser.parse_args() lowercase : Optional[int] = torch.load(args.checkpoint_path) with open(args.config_file) as f: lowercase : Tuple = json.loads(f.read()) lowercase : Optional[Any] = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] lowercase : int = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: lowercase : Tuple = DDPMScheduler.from_config("""/""".join(args.checkpoint_path.split("""/""")[:-1])) lowercase : List[Any] = VQModel.from_pretrained("""/""".join(args.checkpoint_path.split("""/""")[:-1])) lowercase : int = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
99
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: # Base Case if curr_ind == len(UpperCAmelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(UpperCAmelCase ) ): if valid_connection(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # Insert current vertex into path as next transition snake_case_ = next_ver # Validate created path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , curr_ind + 1 ): return True # Backtrack snake_case_ = -1 return False def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = 0 ) -> list[int]: snake_case_ = [-1] * (len(UpperCAmelCase ) + 1) # initialize start and end of path with starting index snake_case_ = snake_case_ = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCAmelCase , UpperCAmelCase , 1 ) else []
69
0
A_ : List[Any] = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A_ : Optional[int] = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A_ : List[Any] = { 0: 'Sunday', 1: 'Monday', 2: 'Tuesday', 3: 'Wednesday', 4: 'Thursday', 5: 'Friday', 6: 'Saturday', } def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' assert len(str(SCREAMING_SNAKE_CASE ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 1_2, "month should be between 1 to 12" assert 1 <= day <= 3_1, "day should be between 1 to 31" # Doomsday algorithm: __UpperCAmelCase = year // 1_0_0 __UpperCAmelCase = (5 * (century % 4) + 2) % 7 __UpperCAmelCase = year % 1_0_0 __UpperCAmelCase = centurian % 1_2 __UpperCAmelCase = ( (centurian // 1_2) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 __UpperCAmelCase = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_0_0) == 0) else DOOMSDAY_LEAP[month - 1] ) __UpperCAmelCase = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
333
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''▁''' __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model''' ), } } __UpperCamelCase = { '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__ = None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = {} if sp_model_kwargs is None else sp_model_kwargs snake_case_ = legacy_behaviour super().__init__( bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, sp_model_kwargs=self.sp_model_kwargs, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowerCAmelCase__)) snake_case_ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token snake_case_ = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab snake_case_ = 1 snake_case_ = len(self.sp_model) snake_case_ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase__) } snake_case_ = {v: k for k, v in self.lang_code_to_id.items()} snake_case_ = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) snake_case_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} snake_case_ = list(self.lang_code_to_id.keys()) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens]) snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.lang_code_to_id[self._src_lang] snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self) -> Union[str, Any]: snake_case_ = self.__dict__.copy() snake_case_ = None snake_case_ = self.sp_model.serialized_model_proto() return state def __setstate__( self, lowerCAmelCase__) -> Tuple: snake_case_ = d # for backward compatibility if not hasattr(self, 'sp_model_kwargs'): snake_case_ = {} snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) @property def a_ ( self) -> str: return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None, lowerCAmelCase__ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__, token_ids_a=lowerCAmelCase__, already_has_special_tokens=lowerCAmelCase__) snake_case_ = [1] * len(self.prefix_tokens) snake_case_ = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__)) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__)) + ([0] * len(lowerCAmelCase__)) + suffix_ones def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self) -> List[Any]: snake_case_ = {self.convert_ids_to_tokens(lowerCAmelCase__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def a_ ( self, lowerCAmelCase__) -> List[str]: return self.sp_model.encode(lowerCAmelCase__, out_type=lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> Any: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case_ = self.sp_model.PieceToId(lowerCAmelCase__) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a_ ( self, lowerCAmelCase__) -> Dict: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def a_ ( self, lowerCAmelCase__) -> List[str]: snake_case_ = ''.join(lowerCAmelCase__).replace(lowerCAmelCase__, ' ').strip() return out_string def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file, lowerCAmelCase__) elif not os.path.isfile(self.vocab_file): with open(lowerCAmelCase__, 'wb') as fi: snake_case_ = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__) return (out_vocab_file,) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> Union[str, Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> int: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[src_lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.lang_code_to_id[lang] if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id]
69
0
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class a (lowerCAmelCase__ ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Union[str, Any] = None , lowerCamelCase : Optional[Any] = None , lowerCamelCase : Optional[Any] = False , **lowerCamelCase : Union[str, Any] , ) -> Dict: super().__init__(features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , keep_in_memory=lowerCAmelCase__ , **lowerCAmelCase__ ) __snake_case : Any = Sql( cache_dir=lowerCAmelCase__ , features=lowerCAmelCase__ , sql=lowerCAmelCase__ , con=lowerCAmelCase__ , **lowerCAmelCase__ , ) def __snake_case ( self : str ) -> int: __snake_case : str = None __snake_case : Dict = None __snake_case : Any = None __snake_case : List[Any] = None self.builder.download_and_prepare( download_config=lowerCAmelCase__ , download_mode=lowerCAmelCase__ , verification_mode=lowerCAmelCase__ , base_path=lowerCAmelCase__ , ) # Build dataset for splits __snake_case : Optional[int] = self.builder.as_dataset( split="train" , verification_mode=lowerCAmelCase__ , in_memory=self.keep_in_memory ) return dataset class a : """simple docstring""" def __init__( self : str , lowerCamelCase : Union[str, Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : List[Any] , lowerCamelCase : Tuple = None , lowerCamelCase : Any = None , **lowerCamelCase : Optional[int] , ) -> Optional[Any]: if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) __snake_case : Optional[Any] = dataset __snake_case : Optional[int] = name __snake_case : int = con __snake_case : Optional[Any] = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __snake_case : List[Any] = num_proc __snake_case : List[Any] = to_sql_kwargs def __snake_case ( self : Optional[int] ) -> int: __snake_case : str = self.to_sql_kwargs.pop("sql" , lowerCAmelCase__ ) __snake_case : Union[str, Any] = self.to_sql_kwargs.pop("con" , lowerCAmelCase__ ) __snake_case : str = self.to_sql_kwargs.pop("index" , lowerCAmelCase__ ) __snake_case : Tuple = self._write(index=lowerCAmelCase__ , **self.to_sql_kwargs ) return written def __snake_case ( self : Optional[Any] , lowerCamelCase : Optional[int] ) -> Any: __snake_case , __snake_case , __snake_case : Optional[int] = args __snake_case : Union[str, Any] = {**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs __snake_case : List[Any] = query_table( table=self.dataset.data , key=slice(lowerCAmelCase__ , offset + self.batch_size ) , indices=self.dataset._indices , ) __snake_case : Optional[int] = batch.to_pandas() __snake_case : Union[str, Any] = df.to_sql(self.name , self.con , index=lowerCAmelCase__ , **lowerCAmelCase__ ) return num_rows or len(lowerCAmelCase__ ) def __snake_case ( self : Any , lowerCamelCase : Optional[int] , **lowerCamelCase : int ) -> int: __snake_case : Optional[Any] = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += self._batch_sql((offset, index, to_sql_kwargs) ) else: __snake_case , __snake_case : str = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql , [(offset, index, to_sql_kwargs) for offset in range(0 , lowerCAmelCase__ , lowerCAmelCase__ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += num_rows return written
123
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCAmelCase ( ) -> int: snake_case_ = HfArgumentParser(UpperCAmelCase ) snake_case_ = parser.parse_args_into_dataclasses()[0] snake_case_ = TensorFlowBenchmark(args=UpperCAmelCase ) try: snake_case_ = parser.parse_args_into_dataclasses()[0] except ValueError as e: snake_case_ = 'Arg --no_{0} is no longer used, please use --no-{0} instead.' snake_case_ = ' '.join(str(UpperCAmelCase ).split(' ' )[:-1] ) snake_case_ = '' snake_case_ = eval(str(UpperCAmelCase ).split(' ' )[-1] ) snake_case_ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = full_error_msg + begin_error_msg + str(UpperCAmelCase ) raise ValueError(UpperCAmelCase ) benchmark.run() if __name__ == "__main__": main()
69
0
def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: while a != 0: lowerCamelCase , lowerCamelCase : Dict = b % a, a return b def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: if gcd(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) != 1: lowerCamelCase : List[str] = f'''mod inverse of {a!r} and {m!r} does not exist''' raise ValueError(_SCREAMING_SNAKE_CASE ) lowerCamelCase , lowerCamelCase , lowerCamelCase : Dict = 1, 0, a lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = 0, 1, m while va != 0: lowerCamelCase : int = ua // va lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[str] = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
48
"""simple docstring""" from __future__ import annotations def UpperCAmelCase ( UpperCAmelCase ) -> None: create_state_space_tree(UpperCAmelCase , [] , 0 , [0 for i in range(len(UpperCAmelCase ) )] ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) -> None: if index == len(UpperCAmelCase ): print(UpperCAmelCase ) return for i in range(len(UpperCAmelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) snake_case_ = True create_state_space_tree(UpperCAmelCase , UpperCAmelCase , index + 1 , UpperCAmelCase ) current_sequence.pop() snake_case_ = False __UpperCamelCase = [3, 1, 2, 4] generate_all_permutations(sequence) __UpperCamelCase = ["A", "B", "C"] generate_all_permutations(sequence_a)
69
0
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class __lowercase ( lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Union[str, Any] = MvpTokenizer _UpperCAmelCase : List[Any] = MvpTokenizerFast _UpperCAmelCase : Optional[int] = True _UpperCAmelCase : Optional[int] = filter_roberta_detectors def _SCREAMING_SNAKE_CASE ( self : Optional[int]): super().setUp() SCREAMING_SNAKE_CASE_: Any = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] SCREAMING_SNAKE_CASE_: Optional[Any] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__)))) SCREAMING_SNAKE_CASE_: Optional[Any] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] SCREAMING_SNAKE_CASE_: List[Any] = {"unk_token": "<unk>"} SCREAMING_SNAKE_CASE_: List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) SCREAMING_SNAKE_CASE_: Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(lowerCAmelCase__) + "\n") with open(self.merges_file , "w" , encoding="utf-8") as fp: fp.write("\n".join(lowerCAmelCase__)) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , **lowerCAmelCase__ : List[str]): kwargs.update(self.special_tokens_map) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : int , **lowerCAmelCase__ : Optional[Any]): kwargs.update(self.special_tokens_map) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase__ : Optional[int]): return "lower newer", "lower newer" @cached_property def _SCREAMING_SNAKE_CASE ( self : Dict): return MvpTokenizer.from_pretrained("RUCAIBox/mvp") @cached_property def _SCREAMING_SNAKE_CASE ( self : int): return MvpTokenizerFast.from_pretrained("RUCAIBox/mvp") @require_torch def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: str = ["A long paragraph for summarization.", "Another paragraph for summarization."] SCREAMING_SNAKE_CASE_: Optional[int] = [0, 250, 251, 1_7818, 13, 3_9186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE_: Optional[Any] = tokenizer(lowerCAmelCase__ , max_length=len(lowerCAmelCase__) , padding=lowerCAmelCase__ , return_tensors="pt") self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) self.assertEqual((2, 9) , batch.input_ids.shape) self.assertEqual((2, 9) , batch.attention_mask.shape) SCREAMING_SNAKE_CASE_: Dict = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__) # Test that special tokens are reset @require_torch def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_: Union[str, Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE_: Optional[Any] = tokenizer(lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="pt") # check if input_ids are returned and no labels self.assertIn("input_ids" , lowerCAmelCase__) self.assertIn("attention_mask" , lowerCAmelCase__) self.assertNotIn("labels" , lowerCAmelCase__) self.assertNotIn("decoder_attention_mask" , lowerCAmelCase__) @require_torch def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_: Optional[int] = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE_: Dict = tokenizer(text_target=lowerCAmelCase__ , max_length=32 , padding="max_length" , return_tensors="pt") self.assertEqual(32 , targets["input_ids"].shape[1]) @require_torch def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE_: Tuple = tokenizer( ["I am a small frog" * 1024, "I am a small frog"] , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="pt") self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) self.assertEqual(batch.input_ids.shape , (2, 1024)) @require_torch def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: List[Any] = ["A long paragraph for summarization."] SCREAMING_SNAKE_CASE_: Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE_: List[str] = tokenizer(lowerCAmelCase__ , text_target=lowerCAmelCase__ , return_tensors="pt") SCREAMING_SNAKE_CASE_: Any = inputs["input_ids"] SCREAMING_SNAKE_CASE_: Union[str, Any] = inputs["labels"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item()) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item()) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item()) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item()) def _SCREAMING_SNAKE_CASE ( self : int): pass def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): SCREAMING_SNAKE_CASE_: Optional[int] = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ , **lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = self.tokenizer_class.from_pretrained(lowerCAmelCase__ , **lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = "A, <mask> AllenNLP sentence." SCREAMING_SNAKE_CASE_: Optional[int] = tokenizer_r.encode_plus(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = tokenizer_p.encode_plus(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"]) , sum(tokens_p["token_type_ids"])) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"]) / len(tokens_r["attention_mask"]) , sum(tokens_p["attention_mask"]) / len(tokens_p["attention_mask"]) , ) SCREAMING_SNAKE_CASE_: Dict = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"]) SCREAMING_SNAKE_CASE_: Any = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"]) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2]) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2]) self.assertSequenceEqual( lowerCAmelCase__ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"]) self.assertSequenceEqual( lowerCAmelCase__ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"])
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( 'Warning: upper bound of deterministic test is exceeded. ' 'Pass allow_probable=True to allow probabilistic test. ' 'A return value of True indicates a probable prime.' ) # array bounds provided by analysis snake_case_ = [ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] snake_case_ = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(UpperCAmelCase , 1 ): if n < _p: # then we have our last prime to check snake_case_ = primes[:idx] break snake_case_ , snake_case_ = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: snake_case_ = False for r in range(UpperCAmelCase ): snake_case_ = pow(UpperCAmelCase , d * 2**r , UpperCAmelCase ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): snake_case_ = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def UpperCAmelCase ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
69
0
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } lowerCamelCase__ = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : int ): """simple docstring""" for attribute in key.split(""".""" ): __a = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: __a = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: __a = hf_pointer.shape assert hf_shape == value.shape, ( f"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" f" {value.shape} for {full_name}" ) if weight_type == "weight": __a = value elif weight_type == "weight_g": __a = value elif weight_type == "weight_v": __a = value elif weight_type == "bias": __a = value else: __a = value logger.info(f"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : int ): """simple docstring""" __a = [] __a = fairseq_model.state_dict() __a = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight __a = None for name, value in fairseq_dict.items(): __a = False if "conv_layers" in name: load_conv_layer( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == """group""" , ) __a = True elif name.split(""".""" )[0] == "proj": __a = fairseq_model.proj __a = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: __a = True if "*" in mapped_key: __a = name.split(_SCREAMING_SNAKE_CASE )[0].split(""".""" )[-2] __a = mapped_key.replace("""*""" , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: __a = """weight_g""" elif "weight_v" in name: __a = """weight_v""" elif "bias" in name: __a = """bias""" elif "weight" in name: __a = """weight""" else: __a = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(f"Unused weights: {unused_weights}" ) return proj_weight def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" __a = full_name.split("""conv_layers.""" )[-1] __a = name.split(""".""" ) __a = int(items[0] ) __a = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) __a = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) __a = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) __a = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"{full_name} has size {value.shape}, but" f" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) __a = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(_SCREAMING_SNAKE_CASE ) def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : int ): """simple docstring""" __a , __a = emb.weight.shape __a = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE ) __a = emb.weight.data return lin_layer def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" with open(_SCREAMING_SNAKE_CASE , """r""" , encoding="""utf-8""" ) as f: __a = f.readlines() __a = [line.split(""" """ )[0] for line in lines] __a = len(_SCREAMING_SNAKE_CASE ) __a = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(_SCREAMING_SNAKE_CASE , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : int , ): """simple docstring""" __a = WavaVecaConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) __a = SpeechaTextaConfig.from_pretrained( _SCREAMING_SNAKE_CASE , vocab_size=_SCREAMING_SNAKE_CASE , decoder_layers=_SCREAMING_SNAKE_CASE , do_stable_layer_norm=_SCREAMING_SNAKE_CASE ) __a = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE , ) __a , __a , __a = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) __a = model[0].eval() # set weights for wav2vec2 encoder __a = WavaVecaModel(_SCREAMING_SNAKE_CASE ) __a = recursively_load_weights_wavaveca(model.encoder , _SCREAMING_SNAKE_CASE ) __a = SpeechaTextaForCausalLM(_SCREAMING_SNAKE_CASE ) __a , __a = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=_SCREAMING_SNAKE_CASE ) # set output linear layer unexpected_keys.remove("""embed_out""" ) __a = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(f"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(f"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) __a = SpeechEncoderDecoderModel(encoder=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE ) __a = False # add projection layer __a = nn.Parameter(projection_layer.weight ) __a = nn.Parameter(projection_layer.bias ) __a = create_vocab_dict(_SCREAMING_SNAKE_CASE ) with open(os.path.join(_SCREAMING_SNAKE_CASE , """vocab.json""" ) , """w""" ) as fp: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __a = SpeechaTextaTokenizer(os.path.join(_SCREAMING_SNAKE_CASE , """vocab.json""" ) ) tokenizer.save_pretrained(_SCREAMING_SNAKE_CASE ) __a = hf_wavavec.config.to_dict() __a = tokenizer.pad_token_id __a = tokenizer.bos_token_id __a = tokenizer.eos_token_id __a = """speech_to_text_2""" __a = """wav2vec2""" __a = SpeechEncoderDecoderConfig.from_dict(_SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(_SCREAMING_SNAKE_CASE ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-large-lv60""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/s2t-small-mustc-en-fr-st""", type=str, help="""Path to hf decoder s2t checkpoint config""", ) parser.add_argument("""--vocab_size""", default=1_0224, type=int, help="""Vocab size of decoder""") parser.add_argument("""--num_decoder_layers""", default=7, type=int, help="""Number of decoder layers""") lowerCamelCase__ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __UpperCamelCase = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
69
0
'''simple docstring''' import argparse A__: Any = '''docs/source/_static/js/custom.js''' def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : str ) -> int: with open(_UpperCAmelCase ,encoding="""utf-8""" ,newline="""\n""" ) as f: _a : Optional[Any] =f.readlines() _a : Dict =0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 _a : Optional[int] =F"const stableVersion = \"v{version}\"\n" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F" \"v{version}\": \"v{version}\",\n" with open(_UpperCAmelCase ,"""w""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: f.writelines(_UpperCAmelCase ) if __name__ == "__main__": A__: List[Any] = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') A__: int = parser.parse_args() update_custom_js(args.version)
276
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: __UpperCamelCase = None __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} __UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } __UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off __UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ = NllbTokenizer SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] def __init__( self, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__="<s>", lowerCAmelCase__="</s>", lowerCAmelCase__="</s>", lowerCAmelCase__="<s>", lowerCAmelCase__="<unk>", lowerCAmelCase__="<pad>", lowerCAmelCase__="<mask>", lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=None, lowerCAmelCase__=False, **lowerCAmelCase__, ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(lowerCAmelCase__, lstrip=lowerCAmelCase__, rstrip=lowerCAmelCase__) if isinstance(lowerCAmelCase__, lowerCAmelCase__) else mask_token snake_case_ = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__, tokenizer_file=lowerCAmelCase__, bos_token=lowerCAmelCase__, eos_token=lowerCAmelCase__, sep_token=lowerCAmelCase__, cls_token=lowerCAmelCase__, unk_token=lowerCAmelCase__, pad_token=lowerCAmelCase__, mask_token=lowerCAmelCase__, src_lang=lowerCAmelCase__, tgt_lang=lowerCAmelCase__, additional_special_tokens=lowerCAmelCase__, legacy_behaviour=lowerCAmelCase__, **lowerCAmelCase__, ) snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True snake_case_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) snake_case_ = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__) for lang_code in FAIRSEQ_LANGUAGE_CODES } snake_case_ = src_lang if src_lang is not None else 'eng_Latn' snake_case_ = self.convert_tokens_to_ids(self._src_lang) snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def a_ ( self) -> str: return self._src_lang @src_lang.setter def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) -> str: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') snake_case_ = src_lang snake_case_ = self(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__, return_tensors=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) snake_case_ = tgt_lang_id return inputs def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = "eng_Latn", lowerCAmelCase__ = None, lowerCAmelCase__ = "fra_Latn", **lowerCAmelCase__, ) -> BatchEncoding: snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__) def a_ ( self) -> List[Any]: return self.set_src_lang_special_tokens(self.src_lang) def a_ ( self) -> Tuple: return self.set_tgt_lang_special_tokens(self.tgt_lang) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__) -> None: snake_case_ = self.convert_tokens_to_ids(lowerCAmelCase__) if self.legacy_behaviour: snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] else: snake_case_ = [self.cur_lang_code] snake_case_ = [self.eos_token_id] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str, pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str, special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str, self.prefix_tokens + self.suffix_tokens)), ) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__ = None) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowerCAmelCase__): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return snake_case_ = os.path.join( lowerCAmelCase__, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCAmelCase__): copyfile(self.vocab_file, lowerCAmelCase__) return (out_vocab_file,)
69
0
'''simple docstring''' import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def _A ( snake_case , snake_case , snake_case ) -> List[str]: _lowercase : Dict = AutoConfig.from_pretrained(snake_case ) _lowercase : Union[str, Any] = FlaxAutoModelForSeqaSeqLM.from_config(config=snake_case ) _lowercase : str = checkpoints.load_tax_checkpoint(snake_case ) _lowercase : Optional[int] = "wi_0" in tax_model["target"]["encoder"]["layers_0"]["mlp"] if config.model_type == "t5": _lowercase : str = "SelfAttention" if config.model_type == "longt5" and config.encoder_attention_type == "local": _lowercase : Optional[Any] = "LocalSelfAttention" elif config.model_type == "longt5" and config.encoder_attention_type == "transient-global": _lowercase : Union[str, Any] = "TransientGlobalSelfAttention" else: raise ValueError( "Given config is expected to have `model_type=\'t5\'`, or `model_type=\'longt5` with `encoder_attention_type`" " attribute with a value from [\'local\', \'transient-global]." ) # Encoder for layer_index in range(config.num_layers ): _lowercase : List[str] = F'''layers_{str(snake_case )}''' # Self-Attention _lowercase : Union[str, Any] = tax_model["target"]["encoder"][layer_name]["attention"]["key"]["kernel"] _lowercase : Union[str, Any] = tax_model["target"]["encoder"][layer_name]["attention"]["out"]["kernel"] _lowercase : Optional[Any] = tax_model["target"]["encoder"][layer_name]["attention"]["query"]["kernel"] _lowercase : Dict = tax_model["target"]["encoder"][layer_name]["attention"]["value"]["kernel"] # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": _lowercase : Any = tax_model["target"]["encoder"][layer_name]["attention"]["T5LayerNorm_0"]["scale"] # Layer Normalization _lowercase : Optional[int] = tax_model["target"]["encoder"][layer_name]["pre_attention_layer_norm"]["scale"] if split_mlp_wi: _lowercase : Optional[int] = tax_model["target"]["encoder"][layer_name]["mlp"]["wi_0"]["kernel"] _lowercase : Optional[Any] = tax_model["target"]["encoder"][layer_name]["mlp"]["wi_1"]["kernel"] else: _lowercase : Optional[int] = tax_model["target"]["encoder"][layer_name]["mlp"]["wi"]["kernel"] _lowercase : Tuple = tax_model["target"]["encoder"][layer_name]["mlp"]["wo"]["kernel"] # Layer Normalization _lowercase : Tuple = tax_model["target"]["encoder"][layer_name]["pre_mlp_layer_norm"]["scale"] # Assigning _lowercase : Union[str, Any] = flax_model.params["encoder"]["block"][str(snake_case )]["layer"] _lowercase : Tuple = tax_attention_key _lowercase : Tuple = tax_attention_out _lowercase : str = tax_attention_query _lowercase : Dict = tax_attention_value _lowercase : int = tax_attention_layer_norm # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": _lowercase : List[str] = tax_global_layer_norm if split_mlp_wi: _lowercase : List[str] = tax_mlp_wi_a _lowercase : List[Any] = tax_mlp_wi_a else: _lowercase : List[Any] = tax_mlp_wi _lowercase : Union[str, Any] = tax_mlp_wo _lowercase : Optional[int] = tax_mlp_layer_norm _lowercase : Tuple = flax_model_encoder_layer_block # Only for layer 0: _lowercase : str = tax_model["target"]["encoder"]["relpos_bias"]["rel_embedding"].T _lowercase : Optional[int] = tax_encoder_rel_embedding # Side/global relative position_bias + layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": _lowercase : Optional[Any] = tax_model["target"]["encoder"]["side_relpos_bias"]["rel_embedding"].T _lowercase : Tuple = tax_encoder_global_rel_embedding # Assigning _lowercase : Optional[Any] = tax_model["target"]["encoder"]["encoder_norm"]["scale"] _lowercase : Tuple = tax_encoder_norm # Decoder for layer_index in range(config.num_layers ): _lowercase : Optional[int] = F'''layers_{str(snake_case )}''' # Self-Attention _lowercase : List[Any] = tax_model["target"]["decoder"][layer_name]["self_attention"]["key"]["kernel"] _lowercase : Dict = tax_model["target"]["decoder"][layer_name]["self_attention"]["out"]["kernel"] _lowercase : List[Any] = tax_model["target"]["decoder"][layer_name]["self_attention"]["query"]["kernel"] _lowercase : Dict = tax_model["target"]["decoder"][layer_name]["self_attention"]["value"]["kernel"] # Layer Normalization _lowercase : Any = tax_model["target"]["decoder"][layer_name]["pre_self_attention_layer_norm"][ "scale" ] # Encoder-Decoder-Attention _lowercase : Optional[Any] = tax_model["target"]["decoder"][layer_name]["encoder_decoder_attention"] _lowercase : Union[str, Any] = tax_enc_dec_attention_module["key"]["kernel"] _lowercase : Optional[Any] = tax_enc_dec_attention_module["out"]["kernel"] _lowercase : str = tax_enc_dec_attention_module["query"]["kernel"] _lowercase : Any = tax_enc_dec_attention_module["value"]["kernel"] # Layer Normalization _lowercase : int = tax_model["target"]["decoder"][layer_name]["pre_cross_attention_layer_norm"]["scale"] # MLP if split_mlp_wi: _lowercase : Optional[int] = tax_model["target"]["decoder"][layer_name]["mlp"]["wi_0"]["kernel"] _lowercase : Tuple = tax_model["target"]["decoder"][layer_name]["mlp"]["wi_1"]["kernel"] else: _lowercase : str = tax_model["target"]["decoder"][layer_name]["mlp"]["wi"]["kernel"] _lowercase : List[str] = tax_model["target"]["decoder"][layer_name]["mlp"]["wo"]["kernel"] # Layer Normalization _lowercase : Optional[int] = tax_model["target"]["decoder"][layer_name]["pre_mlp_layer_norm"]["scale"] # Assigning _lowercase : Tuple = flax_model.params["decoder"]["block"][str(snake_case )]["layer"] _lowercase : List[Any] = tax_attention_key _lowercase : Tuple = tax_attention_out _lowercase : Optional[int] = tax_attention_query _lowercase : Union[str, Any] = tax_attention_value _lowercase : List[str] = tax_pre_attention_layer_norm _lowercase : Tuple = tax_enc_dec_attention_key _lowercase : Any = tax_enc_dec_attention_out _lowercase : List[str] = tax_enc_dec_attention_query _lowercase : int = tax_enc_dec_attention_value _lowercase : List[str] = tax_cross_layer_norm if split_mlp_wi: _lowercase : int = tax_mlp_wi_a _lowercase : Tuple = tax_mlp_wi_a else: _lowercase : Any = tax_mlp_wi _lowercase : Union[str, Any] = tax_mlp_wo _lowercase : Optional[Any] = txa_mlp_layer_norm _lowercase : List[str] = flax_model_decoder_layer_block # Decoder Normalization _lowercase : Any = tax_model["target"]["decoder"]["decoder_norm"]["scale"] _lowercase : Optional[int] = txa_decoder_norm # Only for layer 0: _lowercase : Any = tax_model["target"]["decoder"]["relpos_bias"]["rel_embedding"].T _lowercase : int = tax_decoder_rel_embedding # Token Embeddings _lowercase : List[str] = tax_model["target"]["token_embedder"]["embedding"] _lowercase : Union[str, Any] = txa_token_embeddings # LM Head (only in v1.1 and LongT5 checkpoints) if "logits_dense" in tax_model["target"]["decoder"]: _lowercase : List[str] = tax_model["target"]["decoder"]["logits_dense"]["kernel"] flax_model.save_pretrained(snake_case ) print("T5X Model was sucessfully converted!" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--t5x_checkpoint_path', default=None, type=str, required=True, help='Path the T5X checkpoint.' ) parser.add_argument('--config_name', default=None, type=str, required=True, help='Config name of LongT5/T5 model.') parser.add_argument( '--flax_dump_folder_path', default=None, type=str, required=True, help='Path to the output FLAX model.' ) _snake_case = parser.parse_args() convert_tax_checkpoint_to_flax(args.tax_checkpoint_path, args.config_name, args.flax_dump_folder_path)
250
"""simple docstring""" from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''T5Config''' class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "mt5" SCREAMING_SNAKE_CASE_ = MTaConfig
69
0
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss __a = pytest.mark.integration @require_faiss class A__ ( lowerCAmelCase__ ): """simple docstring""" def _lowerCAmelCase ( self : str ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase : Union[str, Any] = Dataset.from_dict({"filename": ["my_name-train" + "_" + str(lowerCAmelCase__ ) for x in np.arange(3_0 ).tolist()]} ) return dset def _lowerCAmelCase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" import faiss _UpperCAmelCase : Dict = self._create_dummy_dataset() _UpperCAmelCase : Tuple = dset.map( lambda lowerCAmelCase__ , lowerCAmelCase__ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowerCAmelCase__ , keep_in_memory=lowerCAmelCase__ ) _UpperCAmelCase : Tuple = dset.add_faiss_index("vecs" , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) dset.drop_index("vecs" ) def _lowerCAmelCase ( self : int ) -> str: """simple docstring""" import faiss _UpperCAmelCase : Dict = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name="vecs" , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def _lowerCAmelCase ( self : Any ) -> Optional[Any]: """simple docstring""" import faiss _UpperCAmelCase : Optional[int] = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name="vecs" , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase__ ) as tmp_file: dset.save_faiss_index("vecs" , tmp_file.name ) dset.load_faiss_index("vecs2" , tmp_file.name ) os.unlink(tmp_file.name ) _UpperCAmelCase , _UpperCAmelCase : Dict = dset.get_nearest_examples("vecs2" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def _lowerCAmelCase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" _UpperCAmelCase : Optional[Any] = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name="vecs" ) dset.drop_index("vecs" ) self.assertRaises(lowerCAmelCase__ , partial(dset.get_nearest_examples , "vecs2" , np.ones(5 , dtype=np.floataa ) ) ) def _lowerCAmelCase ( self : Optional[Any] ) -> str: """simple docstring""" from elasticsearch import Elasticsearch _UpperCAmelCase : Union[str, Any] = self._create_dummy_dataset() with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: _UpperCAmelCase : List[Any] = {"acknowledged": True} mocked_bulk.return_value([(True, None)] * 3_0 ) _UpperCAmelCase : Optional[Any] = {"hits": {"hits": [{"_score": 1, "_id": 2_9}]}} _UpperCAmelCase : Tuple = Elasticsearch() dset.add_elasticsearch_index("filename" , es_client=lowerCAmelCase__ ) _UpperCAmelCase , _UpperCAmelCase : List[str] = dset.get_nearest_examples("filename" , "my_name-train_29" ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) @require_faiss class A__ ( lowerCAmelCase__ ): """simple docstring""" def _lowerCAmelCase ( self : str ) -> Tuple: """simple docstring""" import faiss _UpperCAmelCase : Dict = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query _UpperCAmelCase : Optional[int] = np.zeros(5 , dtype=np.floataa ) _UpperCAmelCase : List[str] = 1 _UpperCAmelCase , _UpperCAmelCase : List[Any] = index.search(lowerCAmelCase__ ) self.assertRaises(lowerCAmelCase__ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries _UpperCAmelCase : List[Any] = np.eye(5 , dtype=np.floataa )[::-1] _UpperCAmelCase , _UpperCAmelCase : List[str] = index.search_batch(lowerCAmelCase__ ) self.assertRaises(lowerCAmelCase__ , index.search_batch , queries[0] ) _UpperCAmelCase : Dict = [scores[0] for scores in total_scores] _UpperCAmelCase : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase__ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowerCAmelCase__ ) def _lowerCAmelCase ( self : str ) -> Union[str, Any]: """simple docstring""" import faiss _UpperCAmelCase : Optional[int] = FaissIndex(string_factory="Flat" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) _UpperCAmelCase : List[Any] = FaissIndex(string_factory="LSH" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowerCAmelCase__ ): _UpperCAmelCase : Tuple = FaissIndex(string_factory="Flat" , custom_index=faiss.IndexFlat(5 ) ) def _lowerCAmelCase ( self : Optional[int] ) -> List[str]: """simple docstring""" import faiss _UpperCAmelCase : List[str] = faiss.IndexFlat(5 ) _UpperCAmelCase : Union[str, Any] = FaissIndex(custom_index=lowerCAmelCase__ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def _lowerCAmelCase ( self : List[Any] ) -> Any: """simple docstring""" import faiss _UpperCAmelCase : Dict = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase__ ) as tmp_file: index.save(tmp_file.name ) _UpperCAmelCase : List[Any] = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) _UpperCAmelCase : List[str] = np.zeros(5 , dtype=np.floataa ) _UpperCAmelCase : Optional[int] = 1 _UpperCAmelCase , _UpperCAmelCase : List[str] = index.search(lowerCAmelCase__ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def __UpperCAmelCase ( a_: List[Any] ): import faiss _UpperCAmelCase : Dict = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) _UpperCAmelCase : List[Any] = "index.faiss" _UpperCAmelCase : Union[str, Any] = f"""mock://{index_name}""" index.save(a_, storage_options=mockfs.storage_options ) _UpperCAmelCase : int = FaissIndex.load(a_, storage_options=mockfs.storage_options ) _UpperCAmelCase : List[str] = np.zeros(5, dtype=np.floataa ) _UpperCAmelCase : Any = 1 _UpperCAmelCase , _UpperCAmelCase : List[str] = index.search(a_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class A__ ( lowerCAmelCase__ ): """simple docstring""" def _lowerCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" from elasticsearch import Elasticsearch with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: _UpperCAmelCase : int = Elasticsearch() _UpperCAmelCase : List[str] = {"acknowledged": True} _UpperCAmelCase : Optional[Any] = ElasticSearchIndex(es_client=lowerCAmelCase__ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(["foo", "bar", "foobar"] ) # single query _UpperCAmelCase : str = "foo" _UpperCAmelCase : List[Any] = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} _UpperCAmelCase , _UpperCAmelCase : Optional[int] = index.search(lowerCAmelCase__ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout _UpperCAmelCase : Union[str, Any] = "foo" _UpperCAmelCase : int = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} _UpperCAmelCase , _UpperCAmelCase : List[Any] = index.search(lowerCAmelCase__ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries _UpperCAmelCase : Union[str, Any] = ["foo", "bar", "foobar"] _UpperCAmelCase : List[Any] = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} _UpperCAmelCase , _UpperCAmelCase : Any = index.search_batch(lowerCAmelCase__ ) _UpperCAmelCase : List[str] = [scores[0] for scores in total_scores] _UpperCAmelCase : Any = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase__ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase__ ) # batched queries with timeout _UpperCAmelCase : List[str] = ["foo", "bar", "foobar"] _UpperCAmelCase : Optional[Any] = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} _UpperCAmelCase , _UpperCAmelCase : Dict = index.search_batch(lowerCAmelCase__ , request_timeout=3_0 ) _UpperCAmelCase : str = [scores[0] for scores in total_scores] _UpperCAmelCase : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase__ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase__ )
145
"""simple docstring""" import argparse __UpperCamelCase = '''docs/source/_static/js/custom.js''' def UpperCAmelCase ( UpperCAmelCase ) -> int: with open(UpperCAmelCase , encoding='utf-8' , newline='\n' ) as f: snake_case_ = f.readlines() snake_case_ = 0 # First let's put the right version while not lines[index].startswith('const stableVersion =' ): index += 1 snake_case_ = f'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith('const versionMapping = {' ): index += 1 # We go until the end while not lines[index].startswith('}' ): index += 1 # We add the new version at the end lines[index - 1] += f' "v{version}": "v{version}",\n' with open(UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') __UpperCamelCase = parser.parse_args() update_custom_js(args.version)
69
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase_ = {"""configuration_reformer""": ["""REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ReformerConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""ReformerTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""ReformerTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """ReformerAttention""", """ReformerForMaskedLM""", """ReformerForQuestionAnswering""", """ReformerForSequenceClassification""", """ReformerLayer""", """ReformerModel""", """ReformerModelWithLMHead""", """ReformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_reformer import REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ReformerConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer import ReformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer_fast import ReformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_reformer import ( REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ReformerAttention, ReformerForMaskedLM, ReformerForQuestionAnswering, ReformerForSequenceClassification, ReformerLayer, ReformerModel, ReformerModelWithLMHead, ReformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
303
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class UpperCamelCase : def __init__( self, lowerCAmelCase__) -> Optional[int]: snake_case_ = data snake_case_ = None class UpperCamelCase : def __init__( self) -> Dict: snake_case_ = None snake_case_ = None def __iter__( self) -> Iterator[Any]: snake_case_ = self.head while self.head: yield node.data snake_case_ = node.next if node == self.head: break def __len__( self) -> int: return sum(1 for _ in self) def __repr__( self) -> str: return "->".join(str(lowerCAmelCase__) for item in iter(self)) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(len(self), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__) -> None: self.insert_nth(0, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> None: if index < 0 or index > len(self): raise IndexError('list index out of range.') snake_case_ = Node(lowerCAmelCase__) if self.head is None: snake_case_ = new_node # first node points itself snake_case_ = snake_case_ = new_node elif index == 0: # insert at head snake_case_ = self.head snake_case_ = snake_case_ = new_node else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = new_node if index == len(self) - 1: # insert at tail snake_case_ = new_node def a_ ( self) -> str: return self.delete_nth(0) def a_ ( self) -> Any: return self.delete_nth(len(self) - 1) def a_ ( self, lowerCAmelCase__ = 0) -> Any: if not 0 <= index < len(self): raise IndexError('list index out of range.') snake_case_ = self.head if self.head == self.tail: # just one node snake_case_ = snake_case_ = None elif index == 0: # delete head node snake_case_ = self.tail.next.next snake_case_ = self.head.next else: snake_case_ = self.head for _ in range(index - 1): snake_case_ = temp.next snake_case_ = temp.next snake_case_ = temp.next.next if index == len(self) - 1: # delete at tail snake_case_ = temp return delete_node.data def a_ ( self) -> bool: return len(self) == 0 def UpperCAmelCase ( ) -> None: snake_case_ = CircularLinkedList() assert len(UpperCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(UpperCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(UpperCAmelCase ) == i circular_linked_list.insert_nth(UpperCAmelCase , i + 1 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(UpperCAmelCase ) == "->".join(str(UpperCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
69
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _A : Optional[int] = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : int = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys _A : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
202
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCamelCase = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __UpperCamelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) __UpperCamelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __UpperCamelCase = re.compile(r'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __UpperCamelCase = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def UpperCAmelCase ( UpperCAmelCase ) -> List[Any]: snake_case_ = None # source code of `config_class` snake_case_ = inspect.getsource(UpperCAmelCase ) snake_case_ = _re_checkpoint.findall(UpperCAmelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): snake_case_ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: snake_case_ = ckpt_name break return checkpoint def UpperCAmelCase ( ) -> Union[str, Any]: snake_case_ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ = get_checkpoint_from_config_class(UpperCAmelCase ) snake_case_ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(UpperCAmelCase ) if len(UpperCAmelCase ) > 0: snake_case_ = '\n'.join(sorted(UpperCAmelCase ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
lowercase : Optional[Any] = range(2, 2_0 + 1) lowercase : List[str] = [1_0**k for k in range(ks[-1] + 1)] lowercase : Union[str, Any] = {} def A_ ( A__ , A__ , A__ , A__ ) -> Optional[int]: a__ : Optional[Any] = sum(a_i[j] for j in range(A__ , len(A__ ) ) ) a__ : List[Any] = sum(a_i[j] * base[j] for j in range(min(len(A__ ) , A__ ) ) ) a__ , a__ : Tuple = 0, 0 a__ : List[str] = n - i a__ : Optional[Any] = memo.get(A__ ) if sub_memo is not None: a__ : Tuple = sub_memo.get(A__ ) if jumps is not None and len(A__ ) > 0: # find and make the largest jump without going over a__ : Dict = -1 for _k in range(len(A__ ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: a__ : Union[str, Any] = _k break if max_jump >= 0: a__ , a__ , a__ : Dict = jumps[max_jump] # since the difference between jumps is cached, add c a__ : int = diff + c for j in range(min(A__ , len(A__ ) ) ): a__ , a__ : List[str] = divmod(A__ , 10 ) if new_c > 0: add(A__ , A__ , A__ ) else: a__ : Optional[int] = [] else: a__ : str = {c: []} a__ : Any = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps a__ , a__ : Tuple = next_term(A__ , k - 1 , i + dn , A__ ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead a__ , a__ : Tuple = compute(A__ , A__ , i + dn , A__ ) diff += _diff dn += terms_jumped a__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped a__ : List[str] = 0 while j < len(A__ ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(A__ , (diff, dn, k) ) return (diff, dn) def A_ ( A__ , A__ , A__ , A__ ) -> Optional[Any]: if i >= n: return 0, i if k > len(A__ ): a_i.extend([0 for _ in range(k - len(A__ ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) a__ : List[str] = i a__ , a__ , a__ : Union[str, Any] = 0, 0, 0 for j in range(len(A__ ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 a__ : Optional[Any] = ds_c + ds_b diff += addend a__ : List[Any] = 0 for j in range(A__ ): a__ : Optional[int] = a_i[j] + addend a__ , a__ : Optional[int] = divmod(A__ , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(A__ , A__ , A__ ) return diff, i - start_i def A_ ( A__ , A__ , A__ ) -> Union[str, Any]: for j in range(A__ , len(A__ ) ): a__ : Optional[Any] = digits[j] + addend if s >= 10: a__ , a__ : Union[str, Any] = divmod(A__ , 10 ) a__ : Dict = addend // 10 + quotient else: a__ : List[Any] = s a__ : int = addend // 10 if addend == 0: break while addend > 0: a__ , a__ : Dict = divmod(A__ , 10 ) digits.append(A__ ) def A_ ( A__ = 10**15 ) -> int: a__ : List[str] = [1] a__ : int = 1 a__ : str = 0 while True: a__ , a__ : Any = next_term(A__ , 20 , i + dn , A__ ) dn += terms_jumped if dn == n - i: break a__ : Dict = 0 for j in range(len(A__ ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F"""{solution() = }""")
99
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __UpperCamelCase = 0 __UpperCamelCase = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __UpperCamelCase = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __UpperCamelCase = tuple[int, int] class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, ) -> None: snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = g_cost snake_case_ = parent snake_case_ = self.calculate_heuristic() snake_case_ = self.g_cost + self.h_cost def a_ ( self) -> float: snake_case_ = self.pos_x - self.goal_x snake_case_ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase__) + abs(lowerCAmelCase__) else: return sqrt(dy**2 + dx**2) def __lt__( self, lowerCAmelCase__) -> bool: return self.f_cost < other.f_cost class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = Node(start[1], start[0], goal[1], goal[0], 0, lowerCAmelCase__) snake_case_ = Node(goal[1], goal[0], goal[1], goal[0], 9_9999, lowerCAmelCase__) snake_case_ = [self.start] snake_case_ = [] snake_case_ = False def a_ ( self) -> list[TPosition]: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() snake_case_ = self.open_nodes.pop(0) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase__) self.closed_nodes.append(lowerCAmelCase__) snake_case_ = self.get_successors(lowerCAmelCase__) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__) else: self.open_nodes.append(lowerCAmelCase__) return [self.start.pos] def a_ ( self, lowerCAmelCase__) -> list[Node]: snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0]) - 1 and 0 <= pos_y <= len(lowerCAmelCase__) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__, lowerCAmelCase__, self.target.pos_y, self.target.pos_x, parent.g_cost + 1, lowerCAmelCase__, )) return successors def a_ ( self, lowerCAmelCase__) -> list[TPosition]: snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x)) snake_case_ = current_node.parent path.reverse() return path class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__) -> None: snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = AStar(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = False def a_ ( self) -> list[TPosition]: while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() snake_case_ = self.fwd_astar.open_nodes.pop(0) snake_case_ = self.bwd_astar.open_nodes.pop(0) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase__, lowerCAmelCase__) self.fwd_astar.closed_nodes.append(lowerCAmelCase__) self.bwd_astar.closed_nodes.append(lowerCAmelCase__) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase__), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase__), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase__) else: # retrieve the best current path snake_case_ = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase__)) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase__) else: astar.open_nodes.append(lowerCAmelCase__) return [self.fwd_astar.start.pos] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> list[TPosition]: snake_case_ = self.fwd_astar.retrace_path(lowerCAmelCase__) snake_case_ = self.bwd_astar.retrace_path(lowerCAmelCase__) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __UpperCamelCase = (0, 0) __UpperCamelCase = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __UpperCamelCase = time.time() __UpperCamelCase = AStar(init, goal) __UpperCamelCase = a_star.search() __UpperCamelCase = time.time() - start_time print(F"""AStar execution time = {end_time:f} seconds""") __UpperCamelCase = time.time() __UpperCamelCase = BidirectionalAStar(init, goal) __UpperCamelCase = time.time() - bd_start_time print(F"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
69
0
def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> list[str]: '''simple docstring''' return [sentence[i : i + ngram_size] for i in range(len(SCREAMING_SNAKE_CASE ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
333
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: while a != 0: snake_case_ , snake_case_ = b % a, a return b def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: if gcd(UpperCAmelCase , UpperCAmelCase ) != 1: snake_case_ = f'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = 1, 0, a snake_case_ , snake_case_ , snake_case_ = 0, 1, m while va != 0: snake_case_ = ua // va snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
69
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Tuple = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
123
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase = { '''configuration_bert''': ['''BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BertConfig''', '''BertOnnxConfig'''], '''tokenization_bert''': ['''BasicTokenizer''', '''BertTokenizer''', '''WordpieceTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''BertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BertForMaskedLM''', '''BertForMultipleChoice''', '''BertForNextSentencePrediction''', '''BertForPreTraining''', '''BertForQuestionAnswering''', '''BertForSequenceClassification''', '''BertForTokenClassification''', '''BertLayer''', '''BertLMHeadModel''', '''BertModel''', '''BertPreTrainedModel''', '''load_tf_weights_in_bert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBertEmbeddings''', '''TFBertForMaskedLM''', '''TFBertForMultipleChoice''', '''TFBertForNextSentencePrediction''', '''TFBertForPreTraining''', '''TFBertForQuestionAnswering''', '''TFBertForSequenceClassification''', '''TFBertForTokenClassification''', '''TFBertLMHeadModel''', '''TFBertMainLayer''', '''TFBertModel''', '''TFBertPreTrainedModel''', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFBertTokenizer'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FlaxBertForCausalLM''', '''FlaxBertForMaskedLM''', '''FlaxBertForMultipleChoice''', '''FlaxBertForNextSentencePrediction''', '''FlaxBertForPreTraining''', '''FlaxBertForQuestionAnswering''', '''FlaxBertForSequenceClassification''', '''FlaxBertForTokenClassification''', '''FlaxBertModel''', '''FlaxBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar SCREAMING_SNAKE_CASE__ : Optional[int] = TypeVar('T') class UpperCamelCase__ (Generic[T] ): '''simple docstring''' lowerCamelCase_ : Tuple = 4_2 # Cache store of keys lowerCamelCase_ : Tuple = 4_2 # References of the keys in cache lowerCamelCase_ : List[str] = 1_0 # Maximum capacity of cache def __init__( self , UpperCamelCase__ ) -> None: lowerCamelCase : List[Any] = deque() lowerCamelCase : Dict = set() if not n: lowerCamelCase : Optional[Any] = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0." ) else: lowerCamelCase : Tuple = n def _lowercase ( self , UpperCamelCase__ ) -> None: if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: lowerCamelCase : Optional[Any] = self.dq_store.pop() self.key_reference.remove(lowerCAmelCase__ ) else: self.dq_store.remove(lowerCAmelCase__ ) self.dq_store.appendleft(lowerCAmelCase__ ) self.key_reference.add(lowerCAmelCase__ ) def _lowercase ( self ) -> None: for k in self.dq_store: print(lowerCAmelCase__ ) def __repr__( self ) -> str: return F'''LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}''' if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE__ : Optional[Any] = LRUCache(4) lru_cache.refer('A') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('A') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
48
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.linear_k''': '''encoder.layers.*.self_attn.linear_k''', '''self_attn.linear_v''': '''encoder.layers.*.self_attn.linear_v''', '''self_attn.linear_q''': '''encoder.layers.*.self_attn.linear_q''', '''self_attn.pos_bias_u''': '''encoder.layers.*.self_attn.pos_bias_u''', '''self_attn.pos_bias_v''': '''encoder.layers.*.self_attn.pos_bias_v''', '''self_attn.linear_out''': '''encoder.layers.*.self_attn.linear_out''', '''self_attn.linear_pos''': '''encoder.layers.*.self_attn.linear_pos''', '''self_attn.rotary_emb''': '''encoder.embed_positions''', '''self_attn_layer_norm''': '''encoder.layers.*.self_attn_layer_norm''', '''conv_module.pointwise_conv1''': '''encoder.layers.*.conv_module.pointwise_conv1''', '''conv_module.pointwise_conv2''': '''encoder.layers.*.conv_module.pointwise_conv2''', '''conv_module.depthwise_conv''': '''encoder.layers.*.conv_module.depthwise_conv''', '''conv_module.batch_norm''': '''encoder.layers.*.conv_module.batch_norm''', '''conv_module.layer_norm''': '''encoder.layers.*.conv_module.layer_norm''', '''ffn1.w_1''': '''encoder.layers.*.ffn1.intermediate_dense''', '''ffn1.w_2''': '''encoder.layers.*.ffn1.output_dense''', '''ffn1.layer_norm''': '''encoder.layers.*.ffn1_layer_norm''', '''ffn2.w_1''': '''encoder.layers.*.ffn2.intermediate_dense''', '''ffn2.w_2''': '''encoder.layers.*.ffn2.output_dense''', '''ffn2.layer_norm''': '''encoder.layers.*.ffn2_layer_norm''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __UpperCamelCase = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: for attribute in key.split('.' ): snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ) if weight_type is not None: snake_case_ = getattr(UpperCAmelCase , UpperCAmelCase ).shape else: snake_case_ = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": snake_case_ = value elif weight_type == "weight_g": snake_case_ = value elif weight_type == "weight_v": snake_case_ = value elif weight_type == "bias": snake_case_ = value elif weight_type == "running_mean": snake_case_ = value elif weight_type == "running_var": snake_case_ = value elif weight_type == "num_batches_tracked": snake_case_ = value elif weight_type == "inv_freq": snake_case_ = value else: snake_case_ = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = [] snake_case_ = fairseq_model.state_dict() snake_case_ = hf_model.wavaveca_conformer.feature_extractor for name, value in fairseq_dict.items(): snake_case_ = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) snake_case_ = True else: for key, mapped_key in MAPPING.items(): snake_case_ = 'wav2vec2_conformer.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case_ = True if "*" in mapped_key: snake_case_ = name.split(UpperCAmelCase )[0].split('.' )[-2] snake_case_ = mapped_key.replace('*' , UpperCAmelCase ) if "pos_bias_u" in name: snake_case_ = None elif "pos_bias_v" in name: snake_case_ = None elif "weight_g" in name: snake_case_ = 'weight_g' elif "weight_v" in name: snake_case_ = 'weight_v' elif "bias" in name: snake_case_ = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case_ = 'weight' elif "running_mean" in name: snake_case_ = 'running_mean' elif "inv_freq" in name: snake_case_ = 'inv_freq' elif "running_var" in name: snake_case_ = 'running_var' elif "num_batches_tracked" in name: snake_case_ = 'num_batches_tracked' else: snake_case_ = None set_recursively(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) continue if not is_used: unused_weights.append(UpperCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: snake_case_ = full_name.split('conv_layers.' )[-1] snake_case_ = name.split('.' ) snake_case_ = int(items[0] ) snake_case_ = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) snake_case_ = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase ) @torch.no_grad() def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=True ) -> str: if config_path is not None: snake_case_ = WavaVecaConformerConfig.from_pretrained(UpperCAmelCase , hidden_act='swish' ) else: snake_case_ = WavaVecaConformerConfig() if "rope" in checkpoint_path: snake_case_ = 'rotary' if is_finetuned: if dict_path: snake_case_ = Dictionary.load(UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq snake_case_ = target_dict.pad_index snake_case_ = target_dict.bos_index snake_case_ = target_dict.eos_index snake_case_ = len(target_dict.symbols ) snake_case_ = os.path.join(UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase ) ) return os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) snake_case_ = target_dict.indices # fairseq has the <pad> and <s> switched snake_case_ = 0 snake_case_ = 1 with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(UpperCAmelCase , UpperCAmelCase ) snake_case_ = WavaVecaCTCTokenizer( UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase , ) snake_case_ = True if config.feat_extract_norm == 'layer' else False snake_case_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=UpperCAmelCase , return_attention_mask=UpperCAmelCase , ) snake_case_ = WavaVecaProcessor(feature_extractor=UpperCAmelCase , tokenizer=UpperCAmelCase ) processor.save_pretrained(UpperCAmelCase ) snake_case_ = WavaVecaConformerForCTC(UpperCAmelCase ) else: snake_case_ = WavaVecaConformerForPreTraining(UpperCAmelCase ) if is_finetuned: snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: snake_case_ = argparse.Namespace(task='audio_pretraining' ) snake_case_ = fairseq.tasks.setup_task(UpperCAmelCase ) snake_case_ , snake_case_ , snake_case_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=UpperCAmelCase ) snake_case_ = model[0].eval() recursively_load_weights(UpperCAmelCase , UpperCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __UpperCamelCase = parser.parse_args() convert_wavaveca_conformer_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
69
0
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed lowerCAmelCase : Optional[Any] = """true""" def A_ ( _UpperCAmelCase , _UpperCAmelCase=82 , _UpperCAmelCase=16 ): set_seed(42 ) SCREAMING_SNAKE_CASE_: List[str] = RegressionModel() SCREAMING_SNAKE_CASE_: int = deepcopy(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] = RegressionDataset(length=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Any = DataLoader(_UpperCAmelCase , batch_size=_UpperCAmelCase ) model.to(accelerator.device ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = accelerator.prepare(_UpperCAmelCase , _UpperCAmelCase ) return model, ddp_model, dataloader def A_ ( _UpperCAmelCase , _UpperCAmelCase=False ): SCREAMING_SNAKE_CASE_: Optional[int] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) SCREAMING_SNAKE_CASE_: Any = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[Any] = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase ) return outputs with accelerator.main_process_first(): SCREAMING_SNAKE_CASE_: Optional[Any] = dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) SCREAMING_SNAKE_CASE_: int = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(_UpperCAmelCase ): if use_longest: return tokenizer.pad(_UpperCAmelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(_UpperCAmelCase , padding="max_length" , max_length=1_28 , return_tensors="pt" ) return DataLoader(_UpperCAmelCase , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=16 ) def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Any = Accelerator(dispatch_batches=_UpperCAmelCase , split_batches=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = get_dataloader(_UpperCAmelCase , not dispatch_batches ) SCREAMING_SNAKE_CASE_: Tuple = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = accelerator.prepare(_UpperCAmelCase , _UpperCAmelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Union[str, Any] = [] for batch in dataloader: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = batch.values() with torch.no_grad(): SCREAMING_SNAKE_CASE_: List[Any] = model(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Dict = [], [] for logit, targ in logits_and_targets: logits.append(_UpperCAmelCase ) targs.append(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = torch.cat(_UpperCAmelCase ), torch.cat(_UpperCAmelCase ) return logits, targs def A_ ( _UpperCAmelCase , _UpperCAmelCase=82 , _UpperCAmelCase=False , _UpperCAmelCase=False , _UpperCAmelCase=16 ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = get_basic_setup(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = generate_predictions(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) assert ( len(_UpperCAmelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_UpperCAmelCase )}" def A_ ( _UpperCAmelCase = False , _UpperCAmelCase = False ): SCREAMING_SNAKE_CASE_: str = evaluate.load("glue" , "mrpc" ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = get_mrpc_setup(_UpperCAmelCase , _UpperCAmelCase ) # First do baseline SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = setup["no"] model.to(_UpperCAmelCase ) model.eval() for batch in dataloader: batch.to(_UpperCAmelCase ) with torch.inference_mode(): SCREAMING_SNAKE_CASE_: str = model(**_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Any = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_UpperCAmelCase , references=batch["labels"] ) SCREAMING_SNAKE_CASE_: int = metric.compute() # Then do distributed SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): SCREAMING_SNAKE_CASE_: List[str] = model(**_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] = outputs.logits.argmax(dim=-1 ) SCREAMING_SNAKE_CASE_: Union[str, Any] = batch["labels"] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_UpperCAmelCase , references=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def A_ ( ): SCREAMING_SNAKE_CASE_: Any = Accelerator(split_batches=_UpperCAmelCase , dispatch_batches=_UpperCAmelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(_UpperCAmelCase , _UpperCAmelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: SCREAMING_SNAKE_CASE_: Tuple = Accelerator(split_batches=_UpperCAmelCase , dispatch_batches=_UpperCAmelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(_UpperCAmelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) SCREAMING_SNAKE_CASE_: int = Accelerator() test_torch_metrics(_UpperCAmelCase , 5_12 ) accelerator.state._reset_state() def A_ ( _UpperCAmelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase ) -> list: if len(UpperCAmelCase ) <= 1: return [tuple(UpperCAmelCase )] snake_case_ = [] def generate(UpperCAmelCase , UpperCAmelCase ): snake_case_ = [0] * n res.append(tuple(UpperCAmelCase ) ) snake_case_ = 0 while i < n: if c[i] < i: if i % 2 == 0: snake_case_ , snake_case_ = arr[i], arr[0] else: snake_case_ , snake_case_ = arr[i], arr[c[i]] res.append(tuple(UpperCAmelCase ) ) c[i] += 1 snake_case_ = 0 else: snake_case_ = 0 i += 1 generate(len(UpperCAmelCase ) , UpperCAmelCase ) return res if __name__ == "__main__": __UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
69
0
import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): def __init__( self : Optional[int] , __lowercase : Any , __lowercase : Optional[Any] , __lowercase : List[Any] ): '''simple docstring''' __a = dataset __a = process __a = params def __len__( self : Any ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : str , __lowercase : Any ): '''simple docstring''' __a = self.dataset[i] __a = self.process(lowerCAmelCase__ , **self.params ) return processed class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): def __init__( self : Tuple , __lowercase : str , __lowercase : List[str] , __lowercase : Tuple , __lowercase : Union[str, Any]=None ): '''simple docstring''' __a = loader __a = infer __a = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether __a = None __a = loader_batch_size # Internal bookkeeping __a = None __a = None def __len__( self : Dict ): '''simple docstring''' return len(self.loader ) def __iter__( self : Union[str, Any] ): '''simple docstring''' __a = iter(self.loader ) return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice __a = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) __a = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): # Convert ModelOutput to tuple first __a = element.to_tuple() if isinstance(element[0] , torch.Tensor ): __a = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): __a = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): __a = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): __a = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around __a = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers __a = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers __a = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. __a = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 __a = self._loader_batch_data.__class__(lowerCAmelCase__ ) self._loader_batch_index += 1 return result def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch __a = next(self.iterator ) __a = self.infer(lowerCAmelCase__ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCAmelCase__ , torch.Tensor ): __a = processed else: __a = list(processed.keys() )[0] __a = processed[key] if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): __a = len(lowerCAmelCase__ ) else: __a = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. __a = observed_batch_size # Setting internal index to unwrap the batch __a = processed __a = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): def __init__( self : Dict , __lowercase : List[Any] , __lowercase : Optional[Any] , __lowercase : Any , __lowercase : Optional[Any]=None ): '''simple docstring''' super().__init__(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __iter__( self : str ): '''simple docstring''' __a = iter(self.loader ) __a = None return self def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' if self.subiterator is None: __a = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item __a = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators __a = self.infer(next(self.iterator ) , **self.params ) __a = next(self.subiterator ) return processed class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): def __iter__( self : List[str] ): '''simple docstring''' __a = iter(self.loader ) return self def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. __a = False __a = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: __a = self.loader_batch_item() __a = item.pop("""is_last""" ) accumulator.append(lowerCAmelCase__ ) if is_last: return accumulator while not is_last: __a = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(lowerCAmelCase__ , torch.Tensor ): __a = processed else: __a = list(processed.keys() )[0] __a = processed[key] if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): __a = len(lowerCAmelCase__ ) else: __a = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. __a = observed_batch_size __a = processed __a = 0 while self._loader_batch_index < self.loader_batch_size: __a = self.loader_batch_item() __a = item.pop("""is_last""" ) accumulator.append(lowerCAmelCase__ ) if is_last: return accumulator else: __a = processed __a = item.pop("""is_last""" ) accumulator.append(lowerCAmelCase__ ) return accumulator class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): def __init__( self : int , __lowercase : List[str] , __lowercase : List[str] ): '''simple docstring''' __a = dataset __a = key def __len__( self : int ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Optional[int] , __lowercase : Tuple ): '''simple docstring''' return self.dataset[i][self.key] class SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): def __init__( self : Optional[int] , __lowercase : Tuple , __lowercase : str , __lowercase : Optional[Any] ): '''simple docstring''' __a = dataset __a = keya __a = keya def __len__( self : List[Any] ): '''simple docstring''' return len(self.dataset ) def __getitem__( self : Any , __lowercase : str ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
302
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __UpperCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __UpperCamelCase = TaTokenizerFast __UpperCamelCase = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __UpperCamelCase = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
69
0
'''simple docstring''' import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class A__ ( lowerCAmelCase__ ): __UpperCamelCase : List[str] = 42 __UpperCamelCase : Tuple = jnp.floataa __UpperCamelCase : Optional[Any] = True def __UpperCAmelCase ( self :Any ) -> List[str]: '''simple docstring''' super().setup() _a : Tuple =nn.Dense(5 , dtype=self.dtype ) def __call__( self :Dict , *SCREAMING_SNAKE_CASE :Optional[int] , **SCREAMING_SNAKE_CASE :int ) -> Optional[Any]: '''simple docstring''' _a : str =super().__call__(*lowerCAmelCase__ , **lowerCAmelCase__ ) _a : Any =self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class A__ ( lowerCAmelCase__ ): __UpperCamelCase : int = FlaxBigBirdForNaturalQuestionsModule def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Any ,_UpperCAmelCase : Optional[int] ,_UpperCAmelCase : Any ,_UpperCAmelCase : List[Any] ,_UpperCAmelCase : List[str] ,_UpperCAmelCase : Dict ) -> Optional[int]: def cross_entropy(_UpperCAmelCase : Any ,_UpperCAmelCase : Any ,_UpperCAmelCase : List[str]=None ): _a : str =logits.shape[-1] _a : Union[str, Any] =(labels[..., None] == jnp.arange(_UpperCAmelCase )[None]).astype("""f4""" ) _a : str =jax.nn.log_softmax(_UpperCAmelCase ,axis=-1 ) _a : str =-jnp.sum(labels * logits ,axis=-1 ) if reduction is not None: _a : List[Any] =reduction(_UpperCAmelCase ) return loss _a : List[str] =partial(_UpperCAmelCase ,reduction=jnp.mean ) _a : int =cross_entropy(_UpperCAmelCase ,_UpperCAmelCase ) _a : Optional[Any] =cross_entropy(_UpperCAmelCase ,_UpperCAmelCase ) _a : Tuple =cross_entropy(_UpperCAmelCase ,_UpperCAmelCase ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class A__ : __UpperCamelCase : Optional[int] = "google/bigbird-roberta-base" __UpperCamelCase : Tuple = 3000 __UpperCamelCase : str = 1_0500 __UpperCamelCase : str = 128 __UpperCamelCase : Optional[int] = 3 __UpperCamelCase : Dict = 1 __UpperCamelCase : Dict = 5 # tx_args __UpperCamelCase : Optional[Any] = 3e-5 __UpperCamelCase : Dict = 0.0 __UpperCamelCase : int = 2_0000 __UpperCamelCase : List[Any] = 0.0_095 __UpperCamelCase : Any = "bigbird-roberta-natural-questions" __UpperCamelCase : Optional[Any] = "training-expt" __UpperCamelCase : Optional[Any] = "data/nq-training.jsonl" __UpperCamelCase : str = "data/nq-validation.jsonl" def __UpperCAmelCase ( self :Optional[int] ) -> Optional[int]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowerCAmelCase__ ) _a : List[str] =os.path.join(self.base_dir , self.save_dir ) _a : Optional[Any] =self.batch_size_per_device * jax.device_count() @dataclass class A__ : __UpperCamelCase : Union[str, Any] = 42 __UpperCamelCase : Dict = 4096 # no dynamic padding on TPUs def __call__( self :str , SCREAMING_SNAKE_CASE :str ) -> Optional[int]: '''simple docstring''' _a : Any =self.collate_fn(lowerCAmelCase__ ) _a : int =jax.tree_util.tree_map(lowerCAmelCase__ , lowerCAmelCase__ ) return batch def __UpperCAmelCase ( self :Optional[Any] , SCREAMING_SNAKE_CASE :Dict ) -> List[Any]: '''simple docstring''' _a , _a : Tuple =self.fetch_inputs(features["""input_ids"""] ) _a : Union[str, Any] ={ """input_ids""": jnp.array(lowerCAmelCase__ , dtype=jnp.intaa ), """attention_mask""": jnp.array(lowerCAmelCase__ , dtype=jnp.intaa ), """start_labels""": jnp.array(features["""start_token"""] , dtype=jnp.intaa ), """end_labels""": jnp.array(features["""end_token"""] , dtype=jnp.intaa ), """pooled_labels""": jnp.array(features["""category"""] , dtype=jnp.intaa ), } return batch def __UpperCAmelCase ( self :List[Any] , SCREAMING_SNAKE_CASE :List[Any] ) -> Optional[int]: '''simple docstring''' _a : str =[self._fetch_inputs(lowerCAmelCase__ ) for ids in input_ids] return zip(*lowerCAmelCase__ ) def __UpperCAmelCase ( self :Any , SCREAMING_SNAKE_CASE :Optional[int] ) -> Optional[Any]: '''simple docstring''' _a : Optional[int] =[1 for _ in range(len(lowerCAmelCase__ ) )] while len(lowerCAmelCase__ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : str ,_UpperCAmelCase : str ,_UpperCAmelCase : int=None ) -> Dict: if seed is not None: _a : Optional[Any] =dataset.shuffle(seed=_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) // batch_size ): _a : str =dataset[i * batch_size : (i + 1) * batch_size] yield dict(_UpperCAmelCase ) @partial(jax.pmap ,axis_name="""batch""" ) def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Optional[Any] ,_UpperCAmelCase : List[str] ,**_UpperCAmelCase : int ) -> List[str]: def loss_fn(_UpperCAmelCase : Optional[Any] ): _a : List[str] =model_inputs.pop("""start_labels""" ) _a : Tuple =model_inputs.pop("""end_labels""" ) _a : Optional[int] =model_inputs.pop("""pooled_labels""" ) _a : Tuple =state.apply_fn(**_UpperCAmelCase ,params=_UpperCAmelCase ,dropout_rng=_UpperCAmelCase ,train=_UpperCAmelCase ) _a , _a , _a : List[Any] =outputs return state.loss_fn( _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,) _a , _a : List[Any] =jax.random.split(_UpperCAmelCase ) _a : Optional[int] =jax.value_and_grad(_UpperCAmelCase ) _a , _a : List[Any] =grad_fn(state.params ) _a : List[str] =jax.lax.pmean({"""loss""": loss} ,axis_name="""batch""" ) _a : List[Any] =jax.lax.pmean(_UpperCAmelCase ,"""batch""" ) _a : Optional[int] =state.apply_gradients(grads=_UpperCAmelCase ) return state, metrics, new_drp_rng @partial(jax.pmap ,axis_name="""batch""" ) def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Union[str, Any] ,**_UpperCAmelCase : List[str] ) -> List[str]: _a : List[Any] =model_inputs.pop("""start_labels""" ) _a : Optional[int] =model_inputs.pop("""end_labels""" ) _a : str =model_inputs.pop("""pooled_labels""" ) _a : Optional[Any] =state.apply_fn(**_UpperCAmelCase ,params=state.params ,train=_UpperCAmelCase ) _a , _a , _a : str =outputs _a : int =state.loss_fn(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) _a : List[Any] =jax.lax.pmean({"""loss""": loss} ,axis_name="""batch""" ) return metrics class A__ ( train_state.TrainState ): __UpperCamelCase : List[str] = struct.field(pytree_node=lowerCAmelCase__ ) @dataclass class A__ : __UpperCamelCase : int = 42 __UpperCamelCase : Union[str, Any] = 42 __UpperCamelCase : Optional[Any] = 42 __UpperCamelCase : str = 42 __UpperCamelCase : Optional[Any] = 42 __UpperCamelCase : List[Any] = 42 __UpperCamelCase : List[Any] = None def __UpperCAmelCase ( self :List[str] , SCREAMING_SNAKE_CASE :str , SCREAMING_SNAKE_CASE :Union[str, Any] , SCREAMING_SNAKE_CASE :Optional[Any] , SCREAMING_SNAKE_CASE :int=None ) -> Optional[int]: '''simple docstring''' _a : Tuple =model.params _a : int =TrainState.create( apply_fn=model.__call__ , params=lowerCAmelCase__ , tx=lowerCAmelCase__ , loss_fn=lowerCAmelCase__ , ) if ckpt_dir is not None: _a , _a , _a , _a , _a : Tuple =restore_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ ) _a : str ={ """lr""": args.lr, """init_lr""": args.init_lr, """warmup_steps""": args.warmup_steps, """num_train_steps""": num_train_steps, """weight_decay""": args.weight_decay, } _a , _a : str =build_tx(**lowerCAmelCase__ ) _a : Optional[Any] =train_state.TrainState( step=lowerCAmelCase__ , apply_fn=model.__call__ , params=lowerCAmelCase__ , tx=lowerCAmelCase__ , opt_state=lowerCAmelCase__ , ) _a : List[Any] =args _a : Tuple =data_collator _a : int =lr _a : List[Any] =params _a : Dict =jax_utils.replicate(lowerCAmelCase__ ) return state def __UpperCAmelCase ( self :Optional[int] , SCREAMING_SNAKE_CASE :List[Any] , SCREAMING_SNAKE_CASE :List[str] , SCREAMING_SNAKE_CASE :Tuple ) -> Optional[int]: '''simple docstring''' _a : List[Any] =self.args _a : Dict =len(lowerCAmelCase__ ) // args.batch_size _a : Optional[int] =jax.random.PRNGKey(0 ) _a : Any =jax.random.split(lowerCAmelCase__ , jax.device_count() ) for epoch in range(args.max_epochs ): _a : Optional[int] =jnp.array(0 , dtype=jnp.floataa ) _a : List[Any] =get_batched_dataset(lowerCAmelCase__ , args.batch_size , seed=lowerCAmelCase__ ) _a : Optional[int] =0 for batch in tqdm(lowerCAmelCase__ , total=lowerCAmelCase__ , desc=f"Running EPOCH-{epoch}" ): _a : Any =self.data_collator(lowerCAmelCase__ ) _a , _a , _a : int =self.train_step_fn(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 if i % args.logging_steps == 0: _a : List[Any] =jax_utils.unreplicate(state.step ) _a : str =running_loss.item() / i _a : Dict =self.scheduler_fn(state_step - 1 ) _a : Optional[int] =self.evaluate(lowerCAmelCase__ , lowerCAmelCase__ ) _a : Optional[int] ={ """step""": state_step.item(), """eval_loss""": eval_loss.item(), """tr_loss""": tr_loss, """lr""": lr.item(), } tqdm.write(str(lowerCAmelCase__ ) ) self.logger.log(lowerCAmelCase__ , commit=lowerCAmelCase__ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f"-e{epoch}-s{i}" , state=lowerCAmelCase__ ) def __UpperCAmelCase ( self :Union[str, Any] , SCREAMING_SNAKE_CASE :List[str] , SCREAMING_SNAKE_CASE :List[str] ) -> int: '''simple docstring''' _a : List[str] =get_batched_dataset(lowerCAmelCase__ , self.args.batch_size ) _a : List[Any] =len(lowerCAmelCase__ ) // self.args.batch_size _a : str =jnp.array(0 , dtype=jnp.floataa ) _a : Any =0 for batch in tqdm(lowerCAmelCase__ , total=lowerCAmelCase__ , desc="""Evaluating ... """ ): _a : Dict =self.data_collator(lowerCAmelCase__ ) _a : str =self.val_step_fn(lowerCAmelCase__ , **lowerCAmelCase__ ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 return running_loss / i def __UpperCAmelCase ( self :Union[str, Any] , SCREAMING_SNAKE_CASE :List[Any] , SCREAMING_SNAKE_CASE :Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _a : Optional[Any] =jax_utils.unreplicate(lowerCAmelCase__ ) print(f"SAVING CHECKPOINT IN {save_dir}" , end=""" ... """ ) self.model_save_fn(lowerCAmelCase__ , params=state.params ) with open(os.path.join(lowerCAmelCase__ , """opt_state.msgpack""" ) , """wb""" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(lowerCAmelCase__ , """args.joblib""" ) ) joblib.dump(self.data_collator , os.path.join(lowerCAmelCase__ , """data_collator.joblib""" ) ) with open(os.path.join(lowerCAmelCase__ , """training_state.json""" ) , """w""" ) as f: json.dump({"""step""": state.step.item()} , lowerCAmelCase__ ) print("""DONE""" ) def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Tuple ,_UpperCAmelCase : Optional[Any] ) -> Optional[Any]: print(F"RESTORING CHECKPOINT FROM {save_dir}" ,end=""" ... """ ) with open(os.path.join(_UpperCAmelCase ,"""flax_model.msgpack""" ) ,"""rb""" ) as f: _a : Any =from_bytes(state.params ,f.read() ) with open(os.path.join(_UpperCAmelCase ,"""opt_state.msgpack""" ) ,"""rb""" ) as f: _a : Dict =from_bytes(state.opt_state ,f.read() ) _a : Dict =joblib.load(os.path.join(_UpperCAmelCase ,"""args.joblib""" ) ) _a : int =joblib.load(os.path.join(_UpperCAmelCase ,"""data_collator.joblib""" ) ) with open(os.path.join(_UpperCAmelCase ,"""training_state.json""" ) ,"""r""" ) as f: _a : Any =json.load(_UpperCAmelCase ) _a : List[str] =training_state["""step"""] print("""DONE""" ) return params, opt_state, step, args, data_collator def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Optional[Any] ,_UpperCAmelCase : int ,_UpperCAmelCase : List[str] ,_UpperCAmelCase : str ) -> List[str]: _a : str =num_train_steps - warmup_steps _a : str =optax.linear_schedule(init_value=_UpperCAmelCase ,end_value=_UpperCAmelCase ,transition_steps=_UpperCAmelCase ) _a : List[str] =optax.linear_schedule(init_value=_UpperCAmelCase ,end_value=1e-7 ,transition_steps=_UpperCAmelCase ) _a : Any =optax.join_schedules(schedules=[warmup_fn, decay_fn] ,boundaries=[warmup_steps] ) return lr def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Dict ,_UpperCAmelCase : Tuple ,_UpperCAmelCase : Dict ,_UpperCAmelCase : Dict ,_UpperCAmelCase : Any ) -> Optional[int]: def weight_decay_mask(_UpperCAmelCase : str ): _a : str =traverse_util.flatten_dict(_UpperCAmelCase ) _a : List[str] ={k: (v[-1] != """bias""" and v[-2:] != ("""LayerNorm""", """scale""")) for k, v in params.items()} return traverse_util.unflatten_dict(_UpperCAmelCase ) _a : Dict =scheduler_fn(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) _a : Union[str, Any] =optax.adamw(learning_rate=_UpperCAmelCase ,weight_decay=_UpperCAmelCase ,mask=_UpperCAmelCase ) return tx, lr
276
"""simple docstring""" import sys from collections import defaultdict class UpperCamelCase : def __init__( self) -> Optional[int]: snake_case_ = [] def a_ ( self, lowerCAmelCase__) -> Any: return self.node_position[vertex] def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = pos def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> str: if start > size // 2 - 1: return else: if 2 * start + 2 >= size: snake_case_ = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: snake_case_ = 2 * start + 1 else: snake_case_ = 2 * start + 2 if heap[smallest_child] < heap[start]: snake_case_ , snake_case_ = heap[smallest_child], positions[smallest_child] snake_case_ , snake_case_ = ( heap[start], positions[start], ) snake_case_ , snake_case_ = temp, tempa snake_case_ = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child], self.get_position(positions[start])) self.set_position(positions[start], lowerCAmelCase__) self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) -> List[str]: snake_case_ = position[index] while index != 0: snake_case_ = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: snake_case_ = heap[parent] snake_case_ = position[parent] self.set_position(position[parent], lowerCAmelCase__) else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, lowerCAmelCase__) break snake_case_ = parent else: snake_case_ = val snake_case_ = temp self.set_position(lowerCAmelCase__, 0) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = len(lowerCAmelCase__) // 2 - 1 for i in range(lowerCAmelCase__, -1, -1): self.top_to_bottom(lowerCAmelCase__, lowerCAmelCase__, len(lowerCAmelCase__), lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Union[str, Any]: snake_case_ = positions[0] snake_case_ = sys.maxsize self.top_to_bottom(lowerCAmelCase__, 0, len(lowerCAmelCase__), lowerCAmelCase__) return temp def UpperCAmelCase ( UpperCAmelCase ) -> Tuple: snake_case_ = Heap() snake_case_ = [0] * len(UpperCAmelCase ) snake_case_ = [-1] * len(UpperCAmelCase ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph snake_case_ = [] # Heap of Distance of vertices from their neighboring vertex snake_case_ = [] for vertex in range(len(UpperCAmelCase ) ): distance_tv.append(sys.maxsize ) positions.append(UpperCAmelCase ) heap.node_position.append(UpperCAmelCase ) snake_case_ = [] snake_case_ = 1 snake_case_ = sys.maxsize for neighbor, distance in adjacency_list[0]: snake_case_ = 0 snake_case_ = distance heap.heapify(UpperCAmelCase , UpperCAmelCase ) for _ in range(1 , len(UpperCAmelCase ) ): snake_case_ = heap.delete_minimum(UpperCAmelCase , UpperCAmelCase ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) snake_case_ = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(UpperCAmelCase )] ): snake_case_ = distance heap.bottom_to_top( UpperCAmelCase , heap.get_position(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase ) snake_case_ = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > __UpperCamelCase = int(input('''Enter number of edges: ''').strip()) __UpperCamelCase = defaultdict(list) for _ in range(edges_number): __UpperCamelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
69
0
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
250
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase = { '''configuration_mctct''': ['''MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MCTCTConfig'''], '''feature_extraction_mctct''': ['''MCTCTFeatureExtractor'''], '''processing_mctct''': ['''MCTCTProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MCTCTForCTC''', '''MCTCTModel''', '''MCTCTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0
'''simple docstring''' import sys from pathlib import Path __a = Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) __a = {'base': 'patrickvonplaten/wav2vec2_tiny_random', 'robust': 'patrickvonplaten/wav2vec2_tiny_random_robust'} __a = 'zero2' __a = 'zero3' __a = [ZEROa, ZEROa] def __UpperCAmelCase ( a_: Union[str, Any], a_: Optional[int], a_: List[Any] ): # customize the test name generator function as we want both params to appear in the sub-test # name, as by default it shows only the first param _UpperCAmelCase : Tuple = parameterized.to_safe_name("_".join(str(a_ ) for x in param.args ) ) return f"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test __a = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class A__ ( lowerCAmelCase__ ): """simple docstring""" @parameterized.expand(lowerCAmelCase__ , name_func=lowerCAmelCase__ ) def _lowerCAmelCase ( self : int , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str] ) -> int: """simple docstring""" self.run_and_check( stage=lowerCAmelCase__ , model=lowerCAmelCase__ , distributed=lowerCAmelCase__ , fpaa=lowerCAmelCase__ , ) @require_torch_multi_gpu @parameterized.expand(lowerCAmelCase__ , name_func=lowerCAmelCase__ ) def _lowerCAmelCase ( self : List[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : Union[str, Any] ) -> List[str]: """simple docstring""" self.run_and_check( stage=lowerCAmelCase__ , model=lowerCAmelCase__ , distributed=lowerCAmelCase__ , fpaa=lowerCAmelCase__ , ) @parameterized.expand(lowerCAmelCase__ , name_func=lowerCAmelCase__ ) def _lowerCAmelCase ( self : str , lowerCAmelCase__ : Dict , lowerCAmelCase__ : int ) -> Any: """simple docstring""" self.run_and_check( stage=lowerCAmelCase__ , model=lowerCAmelCase__ , distributed=lowerCAmelCase__ , fpaa=lowerCAmelCase__ , ) @require_torch_multi_gpu @parameterized.expand(lowerCAmelCase__ , name_func=lowerCAmelCase__ ) def _lowerCAmelCase ( self : Union[str, Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_and_check( stage=lowerCAmelCase__ , model=lowerCAmelCase__ , distributed=lowerCAmelCase__ , fpaa=lowerCAmelCase__ , ) def _lowerCAmelCase ( self : Optional[int] , lowerCAmelCase__ : Dict ) -> List[Any]: """simple docstring""" pass def _lowerCAmelCase ( self : Tuple , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : Any = 1_0 , lowerCAmelCase__ : Dict = True , lowerCAmelCase__ : Optional[Any] = True , lowerCAmelCase__ : List[str] = True , ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase : List[Any] = models[model] _UpperCAmelCase : Union[str, Any] = self.run_trainer( stage=lowerCAmelCase__ , model_name=lowerCAmelCase__ , eval_steps=lowerCAmelCase__ , num_train_epochs=1 , distributed=lowerCAmelCase__ , fpaa=lowerCAmelCase__ , ) self.do_checks(lowerCAmelCase__ ) return output_dir def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Any = 1_0 , lowerCAmelCase__ : List[str] = 1 , lowerCAmelCase__ : List[Any] = True , lowerCAmelCase__ : Tuple = True , ) -> int: """simple docstring""" _UpperCAmelCase : Any = self.get_auto_remove_tmp_dir("./xxx" , after=lowerCAmelCase__ ) _UpperCAmelCase : List[Any] = F"""\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(lowerCAmelCase__ )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n """.split() if fpaa: args.extend(["--fp16"] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files _UpperCAmelCase : str = F"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() _UpperCAmelCase : Dict = [F"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] _UpperCAmelCase : Dict = self.get_launcher(lowerCAmelCase__ ) _UpperCAmelCase : Optional[int] = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowerCAmelCase__ , env=self.get_env() ) return output_dir def _lowerCAmelCase ( self : Optional[Any] , lowerCAmelCase__ : List[Any]=False ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase : List[Any] = min(2 , get_gpu_count() ) if distributed else 1 return F"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
145
"""simple docstring""" from math import factorial def UpperCAmelCase ( UpperCAmelCase = 20 ) -> int: snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(UpperCAmelCase ) / (factorial(UpperCAmelCase ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: __UpperCamelCase = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
69
0
import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. lowercase_ = abspath(join(dirname(dirname(dirname(__file__))), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def a__ ( snake_case ): """simple docstring""" from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(snake_case ) def a__ ( snake_case ): """simple docstring""" from transformers.testing_utils import pytest_terminal_summary_main __SCREAMING_SNAKE_CASE : Dict = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(snake_case , id=snake_case )
303
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
0